var/home/core/zuul-output/0000755000175000017500000000000015070173540014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070204567015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004706035415070204561017706 0ustar rootrootOct 04 10:34:30 crc systemd[1]: Starting Kubernetes Kubelet... Oct 04 10:34:30 crc restorecon[4693]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:30 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:31 crc restorecon[4693]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 04 10:34:31 crc restorecon[4693]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 04 10:34:32 crc kubenswrapper[5025]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 10:34:32 crc kubenswrapper[5025]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 04 10:34:32 crc kubenswrapper[5025]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 10:34:32 crc kubenswrapper[5025]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 10:34:32 crc kubenswrapper[5025]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 04 10:34:32 crc kubenswrapper[5025]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.092964 5025 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098813 5025 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098850 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098861 5025 feature_gate.go:330] unrecognized feature gate: Example Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098871 5025 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098882 5025 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098892 5025 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098901 5025 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098911 5025 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098920 5025 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098929 5025 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098938 5025 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098946 5025 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098958 5025 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098969 5025 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098979 5025 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098988 5025 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.098997 5025 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099006 5025 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099043 5025 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099052 5025 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099061 5025 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099077 5025 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099085 5025 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099094 5025 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099103 5025 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099112 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099120 5025 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099129 5025 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099138 5025 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099150 5025 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099161 5025 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099171 5025 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099181 5025 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099192 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099201 5025 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099211 5025 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099219 5025 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099228 5025 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099236 5025 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099245 5025 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099253 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099264 5025 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099274 5025 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099282 5025 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099292 5025 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099300 5025 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099309 5025 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099317 5025 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099325 5025 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099334 5025 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099345 5025 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099355 5025 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099363 5025 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099372 5025 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099380 5025 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099389 5025 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099397 5025 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099408 5025 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099416 5025 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099424 5025 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099435 5025 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099448 5025 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099458 5025 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099466 5025 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099475 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099483 5025 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099491 5025 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099500 5025 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099509 5025 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099517 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.099525 5025 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101212 5025 flags.go:64] FLAG: --address="0.0.0.0" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101243 5025 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101261 5025 flags.go:64] FLAG: --anonymous-auth="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101273 5025 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101285 5025 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101295 5025 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101314 5025 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101326 5025 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101367 5025 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101378 5025 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101390 5025 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101401 5025 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101411 5025 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101421 5025 flags.go:64] FLAG: --cgroup-root="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101430 5025 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101440 5025 flags.go:64] FLAG: --client-ca-file="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101449 5025 flags.go:64] FLAG: --cloud-config="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101458 5025 flags.go:64] FLAG: --cloud-provider="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101468 5025 flags.go:64] FLAG: --cluster-dns="[]" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101480 5025 flags.go:64] FLAG: --cluster-domain="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101489 5025 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101499 5025 flags.go:64] FLAG: --config-dir="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101509 5025 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101519 5025 flags.go:64] FLAG: --container-log-max-files="5" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101531 5025 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101541 5025 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101551 5025 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101561 5025 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101571 5025 flags.go:64] FLAG: --contention-profiling="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101582 5025 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101592 5025 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101602 5025 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101612 5025 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101623 5025 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101633 5025 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101643 5025 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101652 5025 flags.go:64] FLAG: --enable-load-reader="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101663 5025 flags.go:64] FLAG: --enable-server="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101673 5025 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101685 5025 flags.go:64] FLAG: --event-burst="100" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101695 5025 flags.go:64] FLAG: --event-qps="50" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101704 5025 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101714 5025 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101724 5025 flags.go:64] FLAG: --eviction-hard="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101735 5025 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101745 5025 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101754 5025 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101764 5025 flags.go:64] FLAG: --eviction-soft="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101774 5025 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101784 5025 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101793 5025 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101803 5025 flags.go:64] FLAG: --experimental-mounter-path="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101812 5025 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101822 5025 flags.go:64] FLAG: --fail-swap-on="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101834 5025 flags.go:64] FLAG: --feature-gates="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101845 5025 flags.go:64] FLAG: --file-check-frequency="20s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101855 5025 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101865 5025 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101875 5025 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101885 5025 flags.go:64] FLAG: --healthz-port="10248" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101895 5025 flags.go:64] FLAG: --help="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101904 5025 flags.go:64] FLAG: --hostname-override="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101913 5025 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101923 5025 flags.go:64] FLAG: --http-check-frequency="20s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101934 5025 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101944 5025 flags.go:64] FLAG: --image-credential-provider-config="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101953 5025 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101962 5025 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101972 5025 flags.go:64] FLAG: --image-service-endpoint="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101981 5025 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.101991 5025 flags.go:64] FLAG: --kube-api-burst="100" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102000 5025 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102010 5025 flags.go:64] FLAG: --kube-api-qps="50" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102051 5025 flags.go:64] FLAG: --kube-reserved="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102061 5025 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102071 5025 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102081 5025 flags.go:64] FLAG: --kubelet-cgroups="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102090 5025 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102100 5025 flags.go:64] FLAG: --lock-file="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102110 5025 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102121 5025 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102131 5025 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102146 5025 flags.go:64] FLAG: --log-json-split-stream="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102156 5025 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102165 5025 flags.go:64] FLAG: --log-text-split-stream="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102175 5025 flags.go:64] FLAG: --logging-format="text" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102186 5025 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102200 5025 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102212 5025 flags.go:64] FLAG: --manifest-url="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102224 5025 flags.go:64] FLAG: --manifest-url-header="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102241 5025 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102253 5025 flags.go:64] FLAG: --max-open-files="1000000" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102268 5025 flags.go:64] FLAG: --max-pods="110" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102280 5025 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102290 5025 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102300 5025 flags.go:64] FLAG: --memory-manager-policy="None" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102310 5025 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102319 5025 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102329 5025 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102339 5025 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102359 5025 flags.go:64] FLAG: --node-status-max-images="50" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102369 5025 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102379 5025 flags.go:64] FLAG: --oom-score-adj="-999" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102389 5025 flags.go:64] FLAG: --pod-cidr="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102398 5025 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102414 5025 flags.go:64] FLAG: --pod-manifest-path="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102423 5025 flags.go:64] FLAG: --pod-max-pids="-1" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102433 5025 flags.go:64] FLAG: --pods-per-core="0" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102443 5025 flags.go:64] FLAG: --port="10250" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102454 5025 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102463 5025 flags.go:64] FLAG: --provider-id="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102473 5025 flags.go:64] FLAG: --qos-reserved="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102483 5025 flags.go:64] FLAG: --read-only-port="10255" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102492 5025 flags.go:64] FLAG: --register-node="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102502 5025 flags.go:64] FLAG: --register-schedulable="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102511 5025 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102528 5025 flags.go:64] FLAG: --registry-burst="10" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102539 5025 flags.go:64] FLAG: --registry-qps="5" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102552 5025 flags.go:64] FLAG: --reserved-cpus="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102563 5025 flags.go:64] FLAG: --reserved-memory="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102578 5025 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102594 5025 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102605 5025 flags.go:64] FLAG: --rotate-certificates="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102615 5025 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102624 5025 flags.go:64] FLAG: --runonce="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102634 5025 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102644 5025 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102654 5025 flags.go:64] FLAG: --seccomp-default="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102664 5025 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102673 5025 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102683 5025 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102692 5025 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102702 5025 flags.go:64] FLAG: --storage-driver-password="root" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102712 5025 flags.go:64] FLAG: --storage-driver-secure="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102721 5025 flags.go:64] FLAG: --storage-driver-table="stats" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102731 5025 flags.go:64] FLAG: --storage-driver-user="root" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102740 5025 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102750 5025 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102760 5025 flags.go:64] FLAG: --system-cgroups="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102769 5025 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102784 5025 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102793 5025 flags.go:64] FLAG: --tls-cert-file="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102802 5025 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102814 5025 flags.go:64] FLAG: --tls-min-version="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102825 5025 flags.go:64] FLAG: --tls-private-key-file="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102837 5025 flags.go:64] FLAG: --topology-manager-policy="none" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102847 5025 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102857 5025 flags.go:64] FLAG: --topology-manager-scope="container" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102866 5025 flags.go:64] FLAG: --v="2" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102878 5025 flags.go:64] FLAG: --version="false" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102890 5025 flags.go:64] FLAG: --vmodule="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102903 5025 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.102917 5025 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103185 5025 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103198 5025 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103208 5025 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103219 5025 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103230 5025 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103240 5025 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103250 5025 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103259 5025 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103267 5025 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103276 5025 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103284 5025 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103293 5025 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103301 5025 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103309 5025 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103317 5025 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103326 5025 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103334 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103342 5025 feature_gate.go:330] unrecognized feature gate: Example Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103350 5025 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103359 5025 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103368 5025 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103376 5025 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103384 5025 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103392 5025 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103400 5025 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103408 5025 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103417 5025 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103425 5025 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103438 5025 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103452 5025 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103464 5025 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103473 5025 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103483 5025 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103492 5025 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103501 5025 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103513 5025 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103525 5025 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103536 5025 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103545 5025 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103555 5025 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103566 5025 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103575 5025 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103585 5025 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103594 5025 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103602 5025 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103611 5025 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103622 5025 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103631 5025 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103639 5025 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103648 5025 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103659 5025 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103671 5025 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103682 5025 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103692 5025 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103701 5025 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103709 5025 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103717 5025 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103726 5025 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103735 5025 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103743 5025 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103751 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103759 5025 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103768 5025 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103776 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103786 5025 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103794 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103803 5025 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103814 5025 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103824 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103833 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.103841 5025 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.103868 5025 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.117108 5025 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.117155 5025 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117280 5025 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117294 5025 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117304 5025 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117316 5025 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117326 5025 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117337 5025 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117386 5025 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117395 5025 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117405 5025 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117414 5025 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117422 5025 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117430 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117439 5025 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117447 5025 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117456 5025 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117464 5025 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117472 5025 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117481 5025 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117490 5025 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117499 5025 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117508 5025 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117517 5025 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117527 5025 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117535 5025 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117544 5025 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117553 5025 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117561 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117569 5025 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117578 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117609 5025 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117620 5025 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117630 5025 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117640 5025 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117650 5025 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117660 5025 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117669 5025 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117677 5025 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117687 5025 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117698 5025 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117711 5025 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117721 5025 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117732 5025 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117744 5025 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117753 5025 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117763 5025 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117774 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117782 5025 feature_gate.go:330] unrecognized feature gate: Example Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117792 5025 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117801 5025 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117810 5025 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117819 5025 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117828 5025 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117837 5025 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117847 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117856 5025 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117866 5025 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117876 5025 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117885 5025 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117893 5025 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117902 5025 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117910 5025 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117919 5025 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117929 5025 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117939 5025 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117949 5025 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117959 5025 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117968 5025 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117977 5025 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117985 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.117993 5025 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118004 5025 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.118047 5025 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118274 5025 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118287 5025 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118296 5025 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118308 5025 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118320 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118330 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118339 5025 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118348 5025 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118358 5025 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118367 5025 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118376 5025 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118387 5025 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118398 5025 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118407 5025 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118416 5025 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118425 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118433 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118444 5025 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118453 5025 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118463 5025 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118471 5025 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118480 5025 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118488 5025 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118496 5025 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118505 5025 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118516 5025 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118527 5025 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118535 5025 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118544 5025 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118553 5025 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118561 5025 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118570 5025 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118579 5025 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118588 5025 feature_gate.go:330] unrecognized feature gate: Example Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118597 5025 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118606 5025 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118615 5025 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118624 5025 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118633 5025 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118641 5025 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118650 5025 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118659 5025 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118667 5025 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118675 5025 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118684 5025 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118692 5025 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118701 5025 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118709 5025 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118720 5025 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118728 5025 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118737 5025 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118745 5025 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118754 5025 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118762 5025 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118770 5025 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118780 5025 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118788 5025 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118797 5025 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118806 5025 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118814 5025 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118822 5025 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118831 5025 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118839 5025 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118848 5025 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118856 5025 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118864 5025 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118873 5025 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118884 5025 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118894 5025 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118904 5025 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.118914 5025 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.118927 5025 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.120258 5025 server.go:940] "Client rotation is on, will bootstrap in background" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.128245 5025 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.128417 5025 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.132107 5025 server.go:997] "Starting client certificate rotation" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.132172 5025 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.133617 5025 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-03 14:37:58.619222644 +0000 UTC Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.133762 5025 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1444h3m26.485465666s for next certificate rotation Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.167049 5025 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.174072 5025 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.204563 5025 log.go:25] "Validated CRI v1 runtime API" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.248661 5025 log.go:25] "Validated CRI v1 image API" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.251926 5025 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.259088 5025 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-04-10-29-21-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.259146 5025 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.291477 5025 manager.go:217] Machine: {Timestamp:2025-10-04 10:34:32.286888845 +0000 UTC m=+0.711855785 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:422640ad-1ab6-4871-9728-16b07446efb8 BootID:707d7ca1-ef0c-4cef-a8ca-1f574b989b45 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:02:1b:c8 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:02:1b:c8 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:5a:d5:da Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:d1:50:93 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:73:d9:77 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ac:22:20 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:86:c4:5c:81:f5:73 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:8e:30:cd:85:dd:4c Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.291890 5025 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.292343 5025 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.295085 5025 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.295519 5025 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.295597 5025 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.295988 5025 topology_manager.go:138] "Creating topology manager with none policy" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.296008 5025 container_manager_linux.go:303] "Creating device plugin manager" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.296782 5025 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.296845 5025 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.297285 5025 state_mem.go:36] "Initialized new in-memory state store" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.297509 5025 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.302480 5025 kubelet.go:418] "Attempting to sync node with API server" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.302523 5025 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.302566 5025 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.302592 5025 kubelet.go:324] "Adding apiserver pod source" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.302615 5025 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.309637 5025 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.318268 5025 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.318900 5025 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.319699 5025 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.318947 5025 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.320102 5025 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.321184 5025 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323120 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323164 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323179 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323194 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323224 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323239 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323254 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323276 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323294 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323310 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323329 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.323342 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.326536 5025 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.327408 5025 server.go:1280] "Started kubelet" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.328805 5025 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.329323 5025 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.329940 5025 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:32 crc systemd[1]: Started Kubernetes Kubelet. Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.330519 5025 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.332355 5025 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.332435 5025 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.332518 5025 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 14:16:05.709864983 +0000 UTC Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.332587 5025 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1203h41m33.377282593s for next certificate rotation Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.333332 5025 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.333370 5025 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.333499 5025 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.334750 5025 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.335085 5025 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.332901 5025 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.339084 5025 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="200ms" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.340695 5025 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.340744 5025 factory.go:55] Registering systemd factory Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.340767 5025 factory.go:221] Registration of the systemd container factory successfully Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.341132 5025 server.go:460] "Adding debug handlers to kubelet server" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.342886 5025 factory.go:153] Registering CRI-O factory Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.342922 5025 factory.go:221] Registration of the crio container factory successfully Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.342966 5025 factory.go:103] Registering Raw factory Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.342997 5025 manager.go:1196] Started watching for new ooms in manager Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.344081 5025 manager.go:319] Starting recovery of all containers Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.341638 5025 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.162:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b434bd10f361e default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-04 10:34:32.327353886 +0000 UTC m=+0.752320796,LastTimestamp:2025-10-04 10:34:32.327353886 +0000 UTC m=+0.752320796,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360233 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360354 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360370 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360386 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360398 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360413 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360426 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360462 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360480 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360494 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360508 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360522 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360537 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360594 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360615 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360630 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360644 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360658 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360672 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360686 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360698 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360710 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360726 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360739 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360755 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360768 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360784 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360799 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360812 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360851 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360865 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360878 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360890 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360902 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360914 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360934 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360947 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360960 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360974 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.360988 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361001 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361031 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361045 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361060 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361076 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361090 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361105 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361119 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361134 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361146 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361159 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361171 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361192 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361208 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361223 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361240 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361255 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361270 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361292 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361314 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361332 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361353 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361370 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361393 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361406 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361420 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361433 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361469 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361483 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361497 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361513 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361525 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361539 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361553 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.361567 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364108 5025 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364164 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364190 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364208 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364234 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364249 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364264 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364331 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364347 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364363 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364377 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364426 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364443 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364457 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364472 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364488 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364505 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364519 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364536 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364550 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364573 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364590 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364607 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364624 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364639 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364655 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364670 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364686 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364702 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364733 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364801 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364829 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364853 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364877 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364895 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364914 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364930 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364946 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364962 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364980 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.364996 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365048 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365065 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365087 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365103 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365116 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365130 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365146 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365161 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365174 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365188 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365203 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365224 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365240 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365255 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365270 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365287 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365303 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365319 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365334 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365380 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365401 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365416 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365432 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365464 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365478 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365493 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365509 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365529 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365545 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365603 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365622 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365639 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365654 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365672 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365690 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365704 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365719 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365734 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365750 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365772 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365789 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365803 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365816 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365834 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365848 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365862 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365877 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365891 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365905 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365917 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365929 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365941 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365956 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365970 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365983 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.365995 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366027 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366041 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366053 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366069 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366084 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366097 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366118 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366131 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366147 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366160 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366178 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366190 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366203 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366214 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366227 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366240 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366253 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366267 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366290 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366306 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366319 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366333 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366347 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366362 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366376 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366389 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366405 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366419 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366437 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366451 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366466 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366481 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366497 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366510 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366523 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366541 5025 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366553 5025 reconstruct.go:97] "Volume reconstruction finished" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.366563 5025 reconciler.go:26] "Reconciler: start to sync state" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.383050 5025 manager.go:324] Recovery completed Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.398195 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.401220 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.401366 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.401389 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.403456 5025 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.403486 5025 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.403532 5025 state_mem.go:36] "Initialized new in-memory state store" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.407505 5025 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.409813 5025 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.409885 5025 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.409974 5025 kubelet.go:2335] "Starting kubelet main sync loop" Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.410147 5025 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.413160 5025 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.413271 5025 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.427441 5025 policy_none.go:49] "None policy: Start" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.431168 5025 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.431223 5025 state_mem.go:35] "Initializing new in-memory state store" Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.438764 5025 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.506741 5025 manager.go:334] "Starting Device Plugin manager" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.507196 5025 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.507236 5025 server.go:79] "Starting device plugin registration server" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.507910 5025 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.507937 5025 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.509124 5025 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.509319 5025 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.509356 5025 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.510261 5025 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.510417 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.514143 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.514245 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.514264 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.514510 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.514808 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.514955 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.516327 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.516391 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.516421 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.516588 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.516731 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.516791 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.516806 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.516843 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.516916 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.518046 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.518104 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.518139 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.518365 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.518385 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.518486 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.518866 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.518925 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.518951 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.519647 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.519672 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.519685 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.519783 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.519820 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.519846 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.520113 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.520823 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.520875 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.521289 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.521341 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.521364 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.521617 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.521641 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.521651 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.521695 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.521752 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.522877 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.522903 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.522916 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.530716 5025 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.540525 5025 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="400ms" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.569446 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.569502 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.569568 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.569598 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.569698 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.569772 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.569802 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.569833 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.569855 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.569962 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.570068 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.570093 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.570117 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.570141 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.570162 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.608500 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.609887 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.609969 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.609993 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.610076 5025 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.610871 5025 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.671903 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.671966 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.671991 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672036 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672079 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672108 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672138 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672163 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672193 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672214 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672241 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672266 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672289 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672311 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672334 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672826 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672871 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672914 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672946 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672954 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672980 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.673001 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.673044 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.673064 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672989 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.673098 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.673101 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.672838 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.673138 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.673164 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.811807 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.814239 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.814300 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.814320 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.814366 5025 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.815041 5025 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.847970 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.856953 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.878601 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.897336 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: I1004 10:34:32.900780 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.901756 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-f33c25f867d4c9b0263789cbf1fc6958cf555d8a954b89119b3102293070d163 WatchSource:0}: Error finding container f33c25f867d4c9b0263789cbf1fc6958cf555d8a954b89119b3102293070d163: Status 404 returned error can't find the container with id f33c25f867d4c9b0263789cbf1fc6958cf555d8a954b89119b3102293070d163 Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.902697 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a85a365188b493bbd984e7347ec336448e07666b4831e700f0819c2064c38d4a WatchSource:0}: Error finding container a85a365188b493bbd984e7347ec336448e07666b4831e700f0819c2064c38d4a: Status 404 returned error can't find the container with id a85a365188b493bbd984e7347ec336448e07666b4831e700f0819c2064c38d4a Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.909893 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-672f66494ef22cbc71bfa885c37f443638bd2d039478f5d5c9c2c3ab20da7122 WatchSource:0}: Error finding container 672f66494ef22cbc71bfa885c37f443638bd2d039478f5d5c9c2c3ab20da7122: Status 404 returned error can't find the container with id 672f66494ef22cbc71bfa885c37f443638bd2d039478f5d5c9c2c3ab20da7122 Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.912947 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-d4b310f4d3facebb5958f95b7ac1d35cb2e3e99cc973faf705d1e00cfb7e1bcc WatchSource:0}: Error finding container d4b310f4d3facebb5958f95b7ac1d35cb2e3e99cc973faf705d1e00cfb7e1bcc: Status 404 returned error can't find the container with id d4b310f4d3facebb5958f95b7ac1d35cb2e3e99cc973faf705d1e00cfb7e1bcc Oct 04 10:34:32 crc kubenswrapper[5025]: W1004 10:34:32.916100 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-6a7276bcadfaec902026ceef9c068477d86da961e2bd8f8675d744ba77eea7a8 WatchSource:0}: Error finding container 6a7276bcadfaec902026ceef9c068477d86da961e2bd8f8675d744ba77eea7a8: Status 404 returned error can't find the container with id 6a7276bcadfaec902026ceef9c068477d86da961e2bd8f8675d744ba77eea7a8 Oct 04 10:34:32 crc kubenswrapper[5025]: E1004 10:34:32.941464 5025 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="800ms" Oct 04 10:34:33 crc kubenswrapper[5025]: W1004 10:34:33.125247 5025 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:33 crc kubenswrapper[5025]: E1004 10:34:33.125348 5025 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.216150 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.217713 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.217763 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.217778 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.217837 5025 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:34:33 crc kubenswrapper[5025]: E1004 10:34:33.218468 5025 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.331702 5025 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.418395 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"672f66494ef22cbc71bfa885c37f443638bd2d039478f5d5c9c2c3ab20da7122"} Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.419518 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a85a365188b493bbd984e7347ec336448e07666b4831e700f0819c2064c38d4a"} Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.420333 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f33c25f867d4c9b0263789cbf1fc6958cf555d8a954b89119b3102293070d163"} Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.421354 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6a7276bcadfaec902026ceef9c068477d86da961e2bd8f8675d744ba77eea7a8"} Oct 04 10:34:33 crc kubenswrapper[5025]: I1004 10:34:33.422081 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d4b310f4d3facebb5958f95b7ac1d35cb2e3e99cc973faf705d1e00cfb7e1bcc"} Oct 04 10:34:33 crc kubenswrapper[5025]: E1004 10:34:33.742740 5025 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="1.6s" Oct 04 10:34:33 crc kubenswrapper[5025]: W1004 10:34:33.789390 5025 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:33 crc kubenswrapper[5025]: E1004 10:34:33.789477 5025 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:34:33 crc kubenswrapper[5025]: W1004 10:34:33.797482 5025 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:33 crc kubenswrapper[5025]: E1004 10:34:33.797556 5025 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:34:33 crc kubenswrapper[5025]: W1004 10:34:33.822752 5025 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:33 crc kubenswrapper[5025]: E1004 10:34:33.822850 5025 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.018985 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.020376 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.020434 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.020446 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.020488 5025 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:34:34 crc kubenswrapper[5025]: E1004 10:34:34.021093 5025 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.331104 5025 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.428427 5025 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f" exitCode=0 Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.428483 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f"} Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.428555 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.429508 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.429543 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.429558 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.431440 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.432059 5025 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1" exitCode=0 Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.432118 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1"} Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.432207 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.432468 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.432510 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.432525 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.432961 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.432995 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.433010 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.436407 5025 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="52644454e1e9f7ec8fb33d08693d42116674144b05f0ce277a92ec07ffa51e5d" exitCode=0 Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.436473 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.436522 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"52644454e1e9f7ec8fb33d08693d42116674144b05f0ce277a92ec07ffa51e5d"} Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.437270 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.437411 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.437516 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.438410 5025 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42" exitCode=0 Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.438498 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42"} Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.438611 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.439489 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.439528 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.439544 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.443275 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd"} Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.443318 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92"} Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.443340 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a"} Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.443360 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11"} Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.443381 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.444178 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.444216 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.444228 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:34 crc kubenswrapper[5025]: I1004 10:34:34.576797 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.331620 5025 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:35 crc kubenswrapper[5025]: E1004 10:34:35.343929 5025 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.162:6443: connect: connection refused" interval="3.2s" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.453455 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029"} Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.453535 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71"} Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.453551 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab"} Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.453569 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5"} Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.456201 5025 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c" exitCode=0 Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.456295 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c"} Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.456475 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.458003 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.458066 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.458081 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.458962 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"01b013405faf0a0ebcab7a507811aaf62fdcec773d83acaf20087fe3a5a58f68"} Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.458969 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.459639 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.459710 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.459724 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.462136 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.462970 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.463167 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b6bc1b4dcad8484ee9c6f82153ed8d93448d8c3540d44e6990981983f9573cbb"} Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.463222 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f018adfdcab8bd60cd03879d0807145bf0e8b398a9a138739387f0f420c7989b"} Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.463237 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cf053ec8fd21d01980e6f7f6d5f24328ff39cf7d641f4ed959ce0ca034f4fbbe"} Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.463761 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.463792 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.463808 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.464123 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.464153 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.464168 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.542711 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.559425 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.621273 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.622846 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.622946 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.622965 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:35 crc kubenswrapper[5025]: I1004 10:34:35.623003 5025 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:34:35 crc kubenswrapper[5025]: E1004 10:34:35.623641 5025 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.162:6443: connect: connection refused" node="crc" Oct 04 10:34:35 crc kubenswrapper[5025]: W1004 10:34:35.740240 5025 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:35 crc kubenswrapper[5025]: E1004 10:34:35.740331 5025 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:34:35 crc kubenswrapper[5025]: W1004 10:34:35.775784 5025 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.162:6443: connect: connection refused Oct 04 10:34:35 crc kubenswrapper[5025]: E1004 10:34:35.775862 5025 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.162:6443: connect: connection refused" logger="UnhandledError" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.466926 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc"} Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.467714 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.468494 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.468573 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.468658 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.470474 5025 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409" exitCode=0 Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.470726 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.470907 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.470678 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.470598 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409"} Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.470853 5025 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.471325 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472055 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472145 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472162 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472170 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472107 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472422 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472513 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472149 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472613 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472080 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472713 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:36 crc kubenswrapper[5025]: I1004 10:34:36.472723 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.476565 5025 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.476622 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.477101 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4"} Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.477136 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462"} Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.477146 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78"} Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.477157 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274"} Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.477291 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.477399 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.477430 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.477440 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.478605 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.478666 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:37 crc kubenswrapper[5025]: I1004 10:34:37.478684 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.324796 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.325588 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.327780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.327871 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.327902 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.486717 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6"} Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.486945 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.488428 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.488500 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.488523 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.824629 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.826865 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.826925 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.826945 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:38 crc kubenswrapper[5025]: I1004 10:34:38.826984 5025 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:34:39 crc kubenswrapper[5025]: I1004 10:34:39.489868 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:39 crc kubenswrapper[5025]: I1004 10:34:39.491408 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:39 crc kubenswrapper[5025]: I1004 10:34:39.491464 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:39 crc kubenswrapper[5025]: I1004 10:34:39.491481 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:39 crc kubenswrapper[5025]: I1004 10:34:39.739522 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:39 crc kubenswrapper[5025]: I1004 10:34:39.739911 5025 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:34:39 crc kubenswrapper[5025]: I1004 10:34:39.739989 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:39 crc kubenswrapper[5025]: I1004 10:34:39.741952 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:39 crc kubenswrapper[5025]: I1004 10:34:39.742063 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:39 crc kubenswrapper[5025]: I1004 10:34:39.742092 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:40 crc kubenswrapper[5025]: I1004 10:34:40.786196 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:40 crc kubenswrapper[5025]: I1004 10:34:40.786499 5025 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:34:40 crc kubenswrapper[5025]: I1004 10:34:40.786567 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:40 crc kubenswrapper[5025]: I1004 10:34:40.788734 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:40 crc kubenswrapper[5025]: I1004 10:34:40.788838 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:40 crc kubenswrapper[5025]: I1004 10:34:40.788860 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:41 crc kubenswrapper[5025]: I1004 10:34:41.768586 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 04 10:34:41 crc kubenswrapper[5025]: I1004 10:34:41.768946 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:41 crc kubenswrapper[5025]: I1004 10:34:41.770919 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:41 crc kubenswrapper[5025]: I1004 10:34:41.770996 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:41 crc kubenswrapper[5025]: I1004 10:34:41.771045 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:41 crc kubenswrapper[5025]: I1004 10:34:41.966338 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:41 crc kubenswrapper[5025]: I1004 10:34:41.966688 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:41 crc kubenswrapper[5025]: I1004 10:34:41.968733 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:41 crc kubenswrapper[5025]: I1004 10:34:41.968928 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:41 crc kubenswrapper[5025]: I1004 10:34:41.968969 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:42 crc kubenswrapper[5025]: I1004 10:34:42.338164 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:42 crc kubenswrapper[5025]: I1004 10:34:42.338465 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:42 crc kubenswrapper[5025]: I1004 10:34:42.340236 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:42 crc kubenswrapper[5025]: I1004 10:34:42.340315 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:42 crc kubenswrapper[5025]: I1004 10:34:42.340335 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:42 crc kubenswrapper[5025]: E1004 10:34:42.531672 5025 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 04 10:34:43 crc kubenswrapper[5025]: I1004 10:34:43.446093 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 04 10:34:43 crc kubenswrapper[5025]: I1004 10:34:43.446394 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:43 crc kubenswrapper[5025]: I1004 10:34:43.448310 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:43 crc kubenswrapper[5025]: I1004 10:34:43.448368 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:43 crc kubenswrapper[5025]: I1004 10:34:43.448388 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:44 crc kubenswrapper[5025]: I1004 10:34:44.420276 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:44 crc kubenswrapper[5025]: I1004 10:34:44.420580 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:44 crc kubenswrapper[5025]: I1004 10:34:44.422655 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:44 crc kubenswrapper[5025]: I1004 10:34:44.422704 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:44 crc kubenswrapper[5025]: I1004 10:34:44.422715 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:44 crc kubenswrapper[5025]: I1004 10:34:44.429231 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:44 crc kubenswrapper[5025]: I1004 10:34:44.504961 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:44 crc kubenswrapper[5025]: I1004 10:34:44.506580 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:44 crc kubenswrapper[5025]: I1004 10:34:44.506623 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:44 crc kubenswrapper[5025]: I1004 10:34:44.506638 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:46 crc kubenswrapper[5025]: I1004 10:34:46.332506 5025 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 04 10:34:46 crc kubenswrapper[5025]: I1004 10:34:46.461406 5025 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 10:34:46 crc kubenswrapper[5025]: I1004 10:34:46.461486 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 10:34:46 crc kubenswrapper[5025]: I1004 10:34:46.467154 5025 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 04 10:34:46 crc kubenswrapper[5025]: I1004 10:34:46.467251 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 04 10:34:47 crc kubenswrapper[5025]: I1004 10:34:47.420345 5025 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 10:34:47 crc kubenswrapper[5025]: I1004 10:34:47.420457 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 04 10:34:49 crc kubenswrapper[5025]: I1004 10:34:49.749319 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:49 crc kubenswrapper[5025]: I1004 10:34:49.749630 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:49 crc kubenswrapper[5025]: I1004 10:34:49.751396 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:49 crc kubenswrapper[5025]: I1004 10:34:49.751511 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:49 crc kubenswrapper[5025]: I1004 10:34:49.751534 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:49 crc kubenswrapper[5025]: I1004 10:34:49.755888 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:50 crc kubenswrapper[5025]: I1004 10:34:50.522071 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:50 crc kubenswrapper[5025]: I1004 10:34:50.523469 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:50 crc kubenswrapper[5025]: I1004 10:34:50.523534 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:50 crc kubenswrapper[5025]: I1004 10:34:50.523554 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:51 crc kubenswrapper[5025]: E1004 10:34:51.463603 5025 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.466244 5025 trace.go:236] Trace[1798981736]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 10:34:39.196) (total time: 12269ms): Oct 04 10:34:51 crc kubenswrapper[5025]: Trace[1798981736]: ---"Objects listed" error: 12269ms (10:34:51.466) Oct 04 10:34:51 crc kubenswrapper[5025]: Trace[1798981736]: [12.269743386s] [12.269743386s] END Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.466270 5025 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.467546 5025 trace.go:236] Trace[299379690]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 10:34:36.756) (total time: 14710ms): Oct 04 10:34:51 crc kubenswrapper[5025]: Trace[299379690]: ---"Objects listed" error: 14710ms (10:34:51.467) Oct 04 10:34:51 crc kubenswrapper[5025]: Trace[299379690]: [14.710601881s] [14.710601881s] END Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.467578 5025 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.467672 5025 trace.go:236] Trace[1720894061]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 10:34:36.944) (total time: 14522ms): Oct 04 10:34:51 crc kubenswrapper[5025]: Trace[1720894061]: ---"Objects listed" error: 14522ms (10:34:51.467) Oct 04 10:34:51 crc kubenswrapper[5025]: Trace[1720894061]: [14.522677931s] [14.522677931s] END Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.467699 5025 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.468558 5025 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 04 10:34:51 crc kubenswrapper[5025]: E1004 10:34:51.469384 5025 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.470158 5025 trace.go:236] Trace[753175616]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Oct-2025 10:34:39.024) (total time: 12445ms): Oct 04 10:34:51 crc kubenswrapper[5025]: Trace[753175616]: ---"Objects listed" error: 12445ms (10:34:51.469) Oct 04 10:34:51 crc kubenswrapper[5025]: Trace[753175616]: [12.445296988s] [12.445296988s] END Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.470638 5025 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.531398 5025 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51096->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.531464 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:51096->192.168.126.11:17697: read: connection reset by peer" Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.531869 5025 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.531932 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.966797 5025 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 04 10:34:51 crc kubenswrapper[5025]: I1004 10:34:51.966895 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.317091 5025 apiserver.go:52] "Watching apiserver" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.322231 5025 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.322468 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.322810 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.322838 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.322882 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.322958 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.322999 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.323092 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.323277 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.323334 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.323361 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.325631 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.325771 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.325960 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.328872 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.329605 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.329813 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.334530 5025 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.338943 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.339239 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.349767 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.375814 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.376168 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.376331 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.376428 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.376658 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.376764 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.376773 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.376777 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.376846 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.376873 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.376959 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377029 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377064 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377089 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377110 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377129 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377149 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377173 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377201 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377224 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377244 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377264 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377285 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377303 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377323 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377345 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377362 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377384 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377402 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377420 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377439 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377462 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377480 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377497 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377513 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377533 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377550 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377567 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377584 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377635 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377651 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377094 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377671 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377214 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377287 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377700 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377424 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377727 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377668 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377763 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377787 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377773 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377817 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377847 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377854 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377881 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377906 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377904 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377931 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377958 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377987 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378038 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378065 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378091 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378119 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378145 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378172 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378200 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378225 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378252 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378283 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378312 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378336 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378361 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378420 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378450 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378483 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378513 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378539 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378565 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378594 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378620 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378646 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378670 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378698 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378727 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378754 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378781 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378806 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378833 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378857 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378882 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378905 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378929 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378954 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378980 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379006 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379050 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379081 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379138 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379165 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379192 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379216 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379242 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379264 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379288 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379312 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379432 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379459 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379499 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379524 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379549 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379578 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379603 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379632 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379657 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379735 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379765 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379789 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379814 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379837 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379863 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379888 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379911 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379937 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379962 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379984 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380027 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380053 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380082 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380105 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380128 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380152 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380178 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380204 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380229 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380258 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380279 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380336 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380359 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380379 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380417 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380444 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380470 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380495 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380522 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380549 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380572 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380599 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380626 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380648 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380671 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380693 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380714 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380735 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380758 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380782 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380805 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380828 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380848 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380872 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380896 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380920 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380944 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380969 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380996 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381120 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381419 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381508 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381562 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381614 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381658 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381693 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381731 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381767 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381808 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381844 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381887 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381927 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.381964 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382001 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382087 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382133 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382171 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382218 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382306 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382344 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382386 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382420 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382456 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382496 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382537 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382574 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382612 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382650 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382692 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382732 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382774 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382811 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382848 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382887 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382926 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382961 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.382999 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383064 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383103 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383141 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383179 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383217 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383257 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383364 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383430 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383474 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383516 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383551 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383589 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383626 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383664 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383706 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383749 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383794 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383836 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383950 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383994 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.377986 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378051 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378067 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378132 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378127 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378187 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378203 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378229 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378289 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378298 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378351 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378377 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378439 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378450 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378555 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378588 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378732 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378726 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378733 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378765 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378879 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378903 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378908 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378919 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378950 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.378979 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379040 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379141 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379154 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379167 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379298 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379337 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379372 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379391 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379517 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379549 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379554 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379587 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379667 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379656 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379761 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379778 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379813 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.379954 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380115 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380214 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.380231 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.383211 5025 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.385978 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.385998 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.385181 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.385355 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.385377 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.385621 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.385721 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.385835 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.386254 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.386642 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.387854 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.388114 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.388313 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.388901 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.389142 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.391640 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.394274 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.394285 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.394462 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.394494 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.394670 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.395140 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.395304 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.397625 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.397720 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.397877 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.398075 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.400534 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.402310 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.402491 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.402745 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.402771 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.403122 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.403330 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.403408 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.403495 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.403481 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.403600 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.403803 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.403825 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.403830 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.404299 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.404388 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.404391 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.404467 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.404560 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.404622 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.404709 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.404843 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.404937 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.405045 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.405221 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.405236 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.405403 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.405651 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.405801 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.406523 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.406747 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.406971 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.407196 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.407493 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.408509 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.408697 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.410214 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.410236 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.410415 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.410544 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.410644 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.410682 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.410555 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.410829 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.410885 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.411091 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.411284 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.411335 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.411360 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.411684 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.411710 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.411952 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.411142 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412051 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412449 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412445 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412468 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412566 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412595 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412640 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412833 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412873 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412944 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412968 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412982 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.413079 5025 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.413186 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:52.913158048 +0000 UTC m=+21.338124928 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.413191 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.413790 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.414335 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.414512 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.414667 5025 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.414934 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.414979 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.415181 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.415220 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.412625 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.415392 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.415310 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.415503 5025 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.416661 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.416953 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.418056 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.418984 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.419134 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.420137 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.427883 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.428064 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.428433 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.428475 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.428527 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.428686 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.428949 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.429037 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.429270 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.429488 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.429537 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.429622 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.429770 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.429845 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:52.929814795 +0000 UTC m=+21.354781675 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.428691 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.430459 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:34:52.930439904 +0000 UTC m=+21.355406794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430489 5025 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430515 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430531 5025 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430546 5025 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430563 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430579 5025 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430593 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430609 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430625 5025 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430640 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430654 5025 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430669 5025 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430670 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430683 5025 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430732 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.430745 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.415655 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.429739 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.430939 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.431226 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.431240 5025 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.431299 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:52.931279689 +0000 UTC m=+21.356246569 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.429840 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.429974 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.437004 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.439821 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.442692 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.442724 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.442742 5025 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.442807 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:52.942783761 +0000 UTC m=+21.367750641 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.442827 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.443215 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.443358 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.443586 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.443923 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.444185 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.444297 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.444554 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.445284 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.445422 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.448862 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.450395 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.451525 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.452352 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.452928 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.455712 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.456382 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.456865 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.459708 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.461623 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.463265 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.464074 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.464849 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.466673 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.467475 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.467816 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.468209 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.469401 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.470146 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.471583 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.472262 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.472863 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.472995 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.474253 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.474884 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.475637 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.476645 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.477467 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.478368 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.478946 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.479645 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.480249 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.480831 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.481860 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.482428 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.482931 5025 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.483469 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.485512 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.486191 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.487107 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.488723 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.489533 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.490717 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.491382 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.491523 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.492889 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.493603 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.494724 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.495431 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.496429 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.496921 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.498090 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.498774 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.500205 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.500795 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.501863 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.502655 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.503537 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.504609 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.504730 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.505236 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.516134 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.526878 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.527903 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.532388 5025 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc" exitCode=255 Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.532444 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc"} Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.532735 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.532898 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.532969 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.532771 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533240 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533267 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533285 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533307 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533320 5025 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533332 5025 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533343 5025 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533357 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533367 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533376 5025 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533386 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533395 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533404 5025 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533414 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533425 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533438 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533449 5025 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533458 5025 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533467 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533476 5025 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533485 5025 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533494 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533503 5025 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533512 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533521 5025 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533530 5025 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533538 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533548 5025 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533557 5025 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533565 5025 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533599 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533608 5025 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533617 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533625 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533634 5025 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533643 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533652 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533662 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533674 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533687 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533720 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533730 5025 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533740 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533748 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533757 5025 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533778 5025 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533791 5025 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533803 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533813 5025 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533822 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533831 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533840 5025 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533851 5025 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533862 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533874 5025 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533887 5025 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533897 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533907 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533918 5025 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533929 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533941 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533953 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533964 5025 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533976 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.533988 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534000 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534030 5025 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534043 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534055 5025 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534065 5025 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534076 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534087 5025 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534099 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534112 5025 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534129 5025 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534141 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534152 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534163 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534175 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534185 5025 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534196 5025 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534208 5025 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534219 5025 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534231 5025 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534241 5025 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534254 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534265 5025 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534276 5025 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534287 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534298 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534315 5025 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534326 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534338 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534350 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534362 5025 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534372 5025 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534385 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534399 5025 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534411 5025 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534425 5025 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534436 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534450 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534463 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534476 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534488 5025 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534501 5025 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534515 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534528 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534540 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534552 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534563 5025 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534576 5025 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534587 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534598 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534634 5025 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534646 5025 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534658 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534669 5025 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534681 5025 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534693 5025 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534705 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534717 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534729 5025 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534741 5025 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534775 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534787 5025 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534798 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534811 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534826 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534840 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534852 5025 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534865 5025 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534879 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534891 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534903 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534916 5025 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534928 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534939 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534951 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534963 5025 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534975 5025 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534987 5025 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.534999 5025 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535028 5025 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535040 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535052 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535063 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535073 5025 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535081 5025 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535090 5025 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535100 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535137 5025 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535146 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535156 5025 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535166 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535176 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535186 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535194 5025 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535204 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535213 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535222 5025 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535231 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535239 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535248 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535317 5025 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535334 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535346 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535357 5025 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535369 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535377 5025 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535386 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535396 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535407 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535416 5025 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535425 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535434 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535444 5025 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535454 5025 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.535462 5025 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.544516 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.549130 5025 scope.go:117] "RemoveContainer" containerID="1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.549335 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.555729 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.566496 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.586559 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.606672 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.619959 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.632642 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.637482 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.642642 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.647489 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.648813 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.657732 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: W1004 10:34:52.658270 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-c49982eaacd0c92439aa96fd14084208dbd27a7b0dbd13f8fdf356f7b311d53a WatchSource:0}: Error finding container c49982eaacd0c92439aa96fd14084208dbd27a7b0dbd13f8fdf356f7b311d53a: Status 404 returned error can't find the container with id c49982eaacd0c92439aa96fd14084208dbd27a7b0dbd13f8fdf356f7b311d53a Oct 04 10:34:52 crc kubenswrapper[5025]: W1004 10:34:52.659623 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-ca6fc3f022186d1c2437ec3916c34e8941560f9881d2f2d354cdb7d65f952efa WatchSource:0}: Error finding container ca6fc3f022186d1c2437ec3916c34e8941560f9881d2f2d354cdb7d65f952efa: Status 404 returned error can't find the container with id ca6fc3f022186d1c2437ec3916c34e8941560f9881d2f2d354cdb7d65f952efa Oct 04 10:34:52 crc kubenswrapper[5025]: W1004 10:34:52.669207 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-5b92311ffe57670d6e0ee6aa21f1616c6352b3a1543655537f064f6136808c3d WatchSource:0}: Error finding container 5b92311ffe57670d6e0ee6aa21f1616c6352b3a1543655537f064f6136808c3d: Status 404 returned error can't find the container with id 5b92311ffe57670d6e0ee6aa21f1616c6352b3a1543655537f064f6136808c3d Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.671988 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.686742 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.702239 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.715186 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.939345 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.939491 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.939545 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:34:53.939513216 +0000 UTC m=+22.364480096 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.939617 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:34:52 crc kubenswrapper[5025]: I1004 10:34:52.939684 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.939624 5025 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.939797 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:53.939776783 +0000 UTC m=+22.364743663 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.939700 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.939820 5025 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.939853 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.939870 5025 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.939877 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:53.939868526 +0000 UTC m=+22.364835626 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:34:52 crc kubenswrapper[5025]: E1004 10:34:52.939912 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:53.939901287 +0000 UTC m=+22.364868367 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.040079 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.040257 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.040279 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.040291 5025 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.040356 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:54.04033547 +0000 UTC m=+22.465302350 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.096636 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-h9kd5"] Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.096965 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-h9kd5" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.098549 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.098669 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.099208 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.110749 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.128651 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.138690 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.141115 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/252f30f4-ff2f-45a1-abbf-da9365c09581-hosts-file\") pod \"node-resolver-h9kd5\" (UID: \"252f30f4-ff2f-45a1-abbf-da9365c09581\") " pod="openshift-dns/node-resolver-h9kd5" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.141161 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c68r2\" (UniqueName: \"kubernetes.io/projected/252f30f4-ff2f-45a1-abbf-da9365c09581-kube-api-access-c68r2\") pod \"node-resolver-h9kd5\" (UID: \"252f30f4-ff2f-45a1-abbf-da9365c09581\") " pod="openshift-dns/node-resolver-h9kd5" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.155226 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.166934 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.190139 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.201952 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.215626 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.242344 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/252f30f4-ff2f-45a1-abbf-da9365c09581-hosts-file\") pod \"node-resolver-h9kd5\" (UID: \"252f30f4-ff2f-45a1-abbf-da9365c09581\") " pod="openshift-dns/node-resolver-h9kd5" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.242388 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c68r2\" (UniqueName: \"kubernetes.io/projected/252f30f4-ff2f-45a1-abbf-da9365c09581-kube-api-access-c68r2\") pod \"node-resolver-h9kd5\" (UID: \"252f30f4-ff2f-45a1-abbf-da9365c09581\") " pod="openshift-dns/node-resolver-h9kd5" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.242513 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/252f30f4-ff2f-45a1-abbf-da9365c09581-hosts-file\") pod \"node-resolver-h9kd5\" (UID: \"252f30f4-ff2f-45a1-abbf-da9365c09581\") " pod="openshift-dns/node-resolver-h9kd5" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.312362 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c68r2\" (UniqueName: \"kubernetes.io/projected/252f30f4-ff2f-45a1-abbf-da9365c09581-kube-api-access-c68r2\") pod \"node-resolver-h9kd5\" (UID: \"252f30f4-ff2f-45a1-abbf-da9365c09581\") " pod="openshift-dns/node-resolver-h9kd5" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.409902 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-h9kd5" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.434673 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-n49fv"] Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.435307 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.435871 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-2dll9"] Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.436292 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-h8qkk"] Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.436718 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.436891 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.441995 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.442250 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.442448 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.442697 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.442756 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.442808 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.442698 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.442700 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.442919 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443086 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443434 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443734 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-system-cni-dir\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443776 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54919b0d-887d-4727-adfc-e48a66e680ba-mcd-auth-proxy-config\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443809 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-var-lib-kubelet\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443831 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-multus-conf-dir\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443849 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8d214492-7532-4f72-b032-b3cc1ae19473-multus-daemon-config\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443867 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-var-lib-cni-bin\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443892 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-multus-socket-dir-parent\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443914 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-os-release\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443931 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8d214492-7532-4f72-b032-b3cc1ae19473-cni-binary-copy\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443952 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-run-k8s-cni-cncf-io\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.443985 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-run-netns\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444062 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-multus-cni-dir\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444082 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-cnibin\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444101 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-etc-kubernetes\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444122 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/54919b0d-887d-4727-adfc-e48a66e680ba-rootfs\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444142 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/54919b0d-887d-4727-adfc-e48a66e680ba-proxy-tls\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444162 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444194 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-hostroot\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444217 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-os-release\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444242 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/24e4becb-78de-4912-ac5f-d25ee8710f40-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444266 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-run-multus-certs\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444287 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/24e4becb-78de-4912-ac5f-d25ee8710f40-cni-binary-copy\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444308 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkmtl\" (UniqueName: \"kubernetes.io/projected/24e4becb-78de-4912-ac5f-d25ee8710f40-kube-api-access-fkmtl\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444331 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95hf8\" (UniqueName: \"kubernetes.io/projected/8d214492-7532-4f72-b032-b3cc1ae19473-kube-api-access-95hf8\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444349 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-cnibin\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444370 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-var-lib-cni-multus\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444392 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zdvw\" (UniqueName: \"kubernetes.io/projected/54919b0d-887d-4727-adfc-e48a66e680ba-kube-api-access-7zdvw\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.444415 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-system-cni-dir\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.449545 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.453716 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.473181 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.484803 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.485256 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.495056 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.501687 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.515972 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.528676 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.538604 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.541970 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544735 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/24e4becb-78de-4912-ac5f-d25ee8710f40-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544784 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-hostroot\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544805 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-os-release\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544824 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkmtl\" (UniqueName: \"kubernetes.io/projected/24e4becb-78de-4912-ac5f-d25ee8710f40-kube-api-access-fkmtl\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544842 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-run-multus-certs\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544869 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/24e4becb-78de-4912-ac5f-d25ee8710f40-cni-binary-copy\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544887 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-cnibin\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544903 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95hf8\" (UniqueName: \"kubernetes.io/projected/8d214492-7532-4f72-b032-b3cc1ae19473-kube-api-access-95hf8\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544920 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-var-lib-cni-multus\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544936 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zdvw\" (UniqueName: \"kubernetes.io/projected/54919b0d-887d-4727-adfc-e48a66e680ba-kube-api-access-7zdvw\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544953 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-system-cni-dir\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.544970 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-system-cni-dir\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545003 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54919b0d-887d-4727-adfc-e48a66e680ba-mcd-auth-proxy-config\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545048 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-var-lib-kubelet\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545066 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-multus-conf-dir\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545134 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8d214492-7532-4f72-b032-b3cc1ae19473-multus-daemon-config\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545174 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-var-lib-cni-bin\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545202 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-multus-socket-dir-parent\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545221 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8d214492-7532-4f72-b032-b3cc1ae19473-cni-binary-copy\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545237 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-run-k8s-cni-cncf-io\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545252 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-run-netns\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545268 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-os-release\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545289 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-multus-cni-dir\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545334 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-cnibin\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545351 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-etc-kubernetes\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545369 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/54919b0d-887d-4727-adfc-e48a66e680ba-rootfs\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545394 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/54919b0d-887d-4727-adfc-e48a66e680ba-proxy-tls\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.545420 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546099 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-tuning-conf-dir\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546176 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-hostroot\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546410 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/24e4becb-78de-4912-ac5f-d25ee8710f40-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546456 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-os-release\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546460 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-multus-conf-dir\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546631 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-etc-kubernetes\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546682 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-var-lib-cni-multus\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546712 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-os-release\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546757 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-run-multus-certs\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546810 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-var-lib-cni-bin\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546863 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-multus-cni-dir\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546879 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-multus-socket-dir-parent\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546905 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-cnibin\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546935 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/54919b0d-887d-4727-adfc-e48a66e680ba-rootfs\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.546965 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-cnibin\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.547093 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-run-k8s-cni-cncf-io\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.547133 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-system-cni-dir\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.547212 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/8d214492-7532-4f72-b032-b3cc1ae19473-multus-daemon-config\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.547256 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/24e4becb-78de-4912-ac5f-d25ee8710f40-system-cni-dir\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.547275 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-var-lib-kubelet\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.547338 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/8d214492-7532-4f72-b032-b3cc1ae19473-host-run-netns\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.547834 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/8d214492-7532-4f72-b032-b3cc1ae19473-cni-binary-copy\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.547979 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/24e4becb-78de-4912-ac5f-d25ee8710f40-cni-binary-copy\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.548173 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54919b0d-887d-4727-adfc-e48a66e680ba-mcd-auth-proxy-config\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.549726 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979"} Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.550441 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.558274 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.566620 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e"} Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.566635 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/54919b0d-887d-4727-adfc-e48a66e680ba-proxy-tls\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.567781 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c"} Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.567829 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ca6fc3f022186d1c2437ec3916c34e8941560f9881d2f2d354cdb7d65f952efa"} Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.570002 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkmtl\" (UniqueName: \"kubernetes.io/projected/24e4becb-78de-4912-ac5f-d25ee8710f40-kube-api-access-fkmtl\") pod \"multus-additional-cni-plugins-h8qkk\" (UID: \"24e4becb-78de-4912-ac5f-d25ee8710f40\") " pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.573003 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.573924 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95hf8\" (UniqueName: \"kubernetes.io/projected/8d214492-7532-4f72-b032-b3cc1ae19473-kube-api-access-95hf8\") pod \"multus-n49fv\" (UID: \"8d214492-7532-4f72-b032-b3cc1ae19473\") " pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.582254 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zdvw\" (UniqueName: \"kubernetes.io/projected/54919b0d-887d-4727-adfc-e48a66e680ba-kube-api-access-7zdvw\") pod \"machine-config-daemon-2dll9\" (UID: \"54919b0d-887d-4727-adfc-e48a66e680ba\") " pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.582353 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494"} Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.582414 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c49982eaacd0c92439aa96fd14084208dbd27a7b0dbd13f8fdf356f7b311d53a"} Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.584781 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-h9kd5" event={"ID":"252f30f4-ff2f-45a1-abbf-da9365c09581","Type":"ContainerStarted","Data":"bc877403f3128a3214f2bf113d5b5bd881466b47985eeede9ab6487e40311a6a"} Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.587534 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5b92311ffe57670d6e0ee6aa21f1616c6352b3a1543655537f064f6136808c3d"} Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.588462 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.599562 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.611850 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.622573 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.634700 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.648339 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.648511 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.663052 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.680776 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.691761 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.706282 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.727574 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.757732 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.784032 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-n49fv" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.790283 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.794438 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.804256 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.811494 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.817272 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bsdl9"] Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.818160 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.822594 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.822641 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.822920 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.822990 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.823106 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.823254 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.823412 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 04 10:34:53 crc kubenswrapper[5025]: W1004 10:34:53.829092 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24e4becb_78de_4912_ac5f_d25ee8710f40.slice/crio-22494665fc5757ec3df19b37e9cc45789673474ceb946fe8d201843893a68ec6 WatchSource:0}: Error finding container 22494665fc5757ec3df19b37e9cc45789673474ceb946fe8d201843893a68ec6: Status 404 returned error can't find the container with id 22494665fc5757ec3df19b37e9cc45789673474ceb946fe8d201843893a68ec6 Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.833387 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.849484 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-systemd-units\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.849657 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-ovn-kubernetes\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.849743 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z75f4\" (UniqueName: \"kubernetes.io/projected/607a1d66-62fc-4dba-9c44-6798f087fb5c-kube-api-access-z75f4\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.849820 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-systemd\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.849897 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-openvswitch\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.850215 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovn-node-metrics-cert\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.850317 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-node-log\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.850387 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.850461 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-ovn\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.850534 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-netns\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.850641 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-netd\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.850776 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-kubelet\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.850871 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-etc-openvswitch\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.851000 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-config\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.851124 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-env-overrides\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.851239 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-bin\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.851360 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-log-socket\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.851475 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-slash\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.851574 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-var-lib-openvswitch\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.851667 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-script-lib\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.859930 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.876043 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.889211 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.904999 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.934794 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.952804 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.952904 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-slash\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.952932 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-var-lib-openvswitch\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.952956 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-script-lib\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.952977 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-systemd-units\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953000 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-ovn-kubernetes\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953039 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z75f4\" (UniqueName: \"kubernetes.io/projected/607a1d66-62fc-4dba-9c44-6798f087fb5c-kube-api-access-z75f4\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953060 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-systemd\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953080 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-openvswitch\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953101 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovn-node-metrics-cert\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953120 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-node-log\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953140 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953161 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-ovn\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953184 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953204 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-netns\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953223 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-netd\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953245 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-etc-openvswitch\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953272 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953296 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-kubelet\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953316 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-config\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953339 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-env-overrides\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953360 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-bin\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953387 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953411 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-log-socket\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953491 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-log-socket\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.953588 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:34:55.953568248 +0000 UTC m=+24.378535128 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953619 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-ovn\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953649 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.953712 5025 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.954996 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:55.95497236 +0000 UTC m=+24.379939240 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953785 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-ovn-kubernetes\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953804 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-openvswitch\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953818 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-netns\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953830 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-slash\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953857 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-netd\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953856 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-var-lib-openvswitch\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953897 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-etc-openvswitch\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.954122 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-systemd\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.954164 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-node-log\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.954196 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-systemd-units\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.954224 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-bin\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.954554 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-script-lib\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.954074 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.954714 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-env-overrides\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.955189 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.955207 5025 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.955255 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:55.955244438 +0000 UTC m=+24.380211318 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.953776 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-kubelet\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.955148 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-config\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.954262 5025 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:34:53 crc kubenswrapper[5025]: E1004 10:34:53.955419 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:55.955408273 +0000 UTC m=+24.380375273 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.958681 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.959611 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovn-node-metrics-cert\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.974056 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.974894 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z75f4\" (UniqueName: \"kubernetes.io/projected/607a1d66-62fc-4dba-9c44-6798f087fb5c-kube-api-access-z75f4\") pod \"ovnkube-node-bsdl9\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:53 crc kubenswrapper[5025]: I1004 10:34:53.992354 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:53Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.019559 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.036368 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.052195 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.054583 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:34:54 crc kubenswrapper[5025]: E1004 10:34:54.054758 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:34:54 crc kubenswrapper[5025]: E1004 10:34:54.054799 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:34:54 crc kubenswrapper[5025]: E1004 10:34:54.054817 5025 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:54 crc kubenswrapper[5025]: E1004 10:34:54.054879 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:56.054859877 +0000 UTC m=+24.479826827 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.071346 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.087537 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.101145 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.114315 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.133165 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.154971 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.161988 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: W1004 10:34:54.168057 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod607a1d66_62fc_4dba_9c44_6798f087fb5c.slice/crio-9e7005ffd447d70030f29b8d24ec07b4986c5d1cde8ec69703fc2a3907044941 WatchSource:0}: Error finding container 9e7005ffd447d70030f29b8d24ec07b4986c5d1cde8ec69703fc2a3907044941: Status 404 returned error can't find the container with id 9e7005ffd447d70030f29b8d24ec07b4986c5d1cde8ec69703fc2a3907044941 Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.176850 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.193330 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.211267 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.230273 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.410953 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.410981 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.411058 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:34:54 crc kubenswrapper[5025]: E1004 10:34:54.411116 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:34:54 crc kubenswrapper[5025]: E1004 10:34:54.411224 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:34:54 crc kubenswrapper[5025]: E1004 10:34:54.411280 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.424088 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.428104 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.459035 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.493671 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.494671 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.509680 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.535059 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.562133 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.591793 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-h9kd5" event={"ID":"252f30f4-ff2f-45a1-abbf-da9365c09581","Type":"ContainerStarted","Data":"f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595"} Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.593819 5025 generic.go:334] "Generic (PLEG): container finished" podID="24e4becb-78de-4912-ac5f-d25ee8710f40" containerID="d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428" exitCode=0 Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.593914 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" event={"ID":"24e4becb-78de-4912-ac5f-d25ee8710f40","Type":"ContainerDied","Data":"d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428"} Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.593980 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" event={"ID":"24e4becb-78de-4912-ac5f-d25ee8710f40","Type":"ContainerStarted","Data":"22494665fc5757ec3df19b37e9cc45789673474ceb946fe8d201843893a68ec6"} Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.597251 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86"} Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.597313 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b"} Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.597325 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"01905916bd7712070da79ebd4a524f200b849d8e2ef178542804f864a65568cd"} Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.598615 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n49fv" event={"ID":"8d214492-7532-4f72-b032-b3cc1ae19473","Type":"ContainerStarted","Data":"507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172"} Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.598644 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n49fv" event={"ID":"8d214492-7532-4f72-b032-b3cc1ae19473","Type":"ContainerStarted","Data":"129f34f9b30378307433950250a8675799906af7cd7805cecefe82b458ec996e"} Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.600137 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495" exitCode=0 Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.600812 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495"} Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.600845 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"9e7005ffd447d70030f29b8d24ec07b4986c5d1cde8ec69703fc2a3907044941"} Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.604263 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: E1004 10:34:54.625962 5025 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.636578 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.656917 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.672691 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.695949 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.711460 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.730452 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.778821 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.810280 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.864507 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.891676 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.931085 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:54 crc kubenswrapper[5025]: I1004 10:34:54.981321 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.014058 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.060492 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.088817 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.129402 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.176070 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.206897 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.256120 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.287540 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.328714 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.610059 5025 generic.go:334] "Generic (PLEG): container finished" podID="24e4becb-78de-4912-ac5f-d25ee8710f40" containerID="6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8" exitCode=0 Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.610483 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" event={"ID":"24e4becb-78de-4912-ac5f-d25ee8710f40","Type":"ContainerDied","Data":"6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8"} Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.614907 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4"} Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.614974 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906"} Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.614988 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95"} Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.615002 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae"} Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.644249 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.658765 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.678308 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.695782 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.712036 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.740255 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.770253 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.791903 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.820438 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.841259 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.859038 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.875673 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.890250 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.902767 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.978644 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:34:55 crc kubenswrapper[5025]: E1004 10:34:55.978895 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:34:59.978860367 +0000 UTC m=+28.403827267 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.978958 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.979577 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:34:55 crc kubenswrapper[5025]: E1004 10:34:55.979612 5025 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:34:55 crc kubenswrapper[5025]: E1004 10:34:55.979750 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:59.979697412 +0000 UTC m=+28.404664292 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:34:55 crc kubenswrapper[5025]: I1004 10:34:55.979926 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:55 crc kubenswrapper[5025]: E1004 10:34:55.982418 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:34:55 crc kubenswrapper[5025]: E1004 10:34:55.982475 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:34:55 crc kubenswrapper[5025]: E1004 10:34:55.982539 5025 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:55 crc kubenswrapper[5025]: E1004 10:34:55.982638 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:59.982598019 +0000 UTC m=+28.407564899 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:55 crc kubenswrapper[5025]: E1004 10:34:55.982787 5025 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:34:55 crc kubenswrapper[5025]: E1004 10:34:55.982871 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:34:59.982852616 +0000 UTC m=+28.407819506 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.081236 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:34:56 crc kubenswrapper[5025]: E1004 10:34:56.081469 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:34:56 crc kubenswrapper[5025]: E1004 10:34:56.081512 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:34:56 crc kubenswrapper[5025]: E1004 10:34:56.081527 5025 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:56 crc kubenswrapper[5025]: E1004 10:34:56.081603 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:00.081583039 +0000 UTC m=+28.506549919 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.411692 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.411772 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.411696 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:34:56 crc kubenswrapper[5025]: E1004 10:34:56.411928 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:34:56 crc kubenswrapper[5025]: E1004 10:34:56.412151 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:34:56 crc kubenswrapper[5025]: E1004 10:34:56.412448 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.625409 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4"} Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.625474 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648"} Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.629484 5025 generic.go:334] "Generic (PLEG): container finished" podID="24e4becb-78de-4912-ac5f-d25ee8710f40" containerID="228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f" exitCode=0 Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.629630 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" event={"ID":"24e4becb-78de-4912-ac5f-d25ee8710f40","Type":"ContainerDied","Data":"228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f"} Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.631555 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0"} Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.651161 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.676831 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.693872 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.723562 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.738510 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.752512 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.767494 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.780832 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.800306 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.816236 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.828938 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.848719 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.863792 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.875479 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.887946 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.901233 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.914821 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.928027 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.952849 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.973398 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.986839 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:56 crc kubenswrapper[5025]: I1004 10:34:56.997077 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.012955 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.027198 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.050242 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.067343 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.084194 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.097079 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.638329 5025 generic.go:334] "Generic (PLEG): container finished" podID="24e4becb-78de-4912-ac5f-d25ee8710f40" containerID="639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3" exitCode=0 Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.638420 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" event={"ID":"24e4becb-78de-4912-ac5f-d25ee8710f40","Type":"ContainerDied","Data":"639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3"} Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.679472 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.697634 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.717336 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.729795 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.743169 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.764385 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.777634 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.789286 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.804626 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.820337 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.836689 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.851745 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.866276 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.869449 5025 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.871511 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.871554 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.871562 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.871684 5025 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.878423 5025 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.878696 5025 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.879899 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.879935 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.879946 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.879965 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.879976 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:57Z","lastTransitionTime":"2025-10-04T10:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.881256 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: E1004 10:34:57.894889 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.900535 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.900579 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.900589 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.900606 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.900617 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:57Z","lastTransitionTime":"2025-10-04T10:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:57 crc kubenswrapper[5025]: E1004 10:34:57.914043 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.918408 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.918464 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.918476 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.918498 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.918516 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:57Z","lastTransitionTime":"2025-10-04T10:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:57 crc kubenswrapper[5025]: E1004 10:34:57.933118 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.937576 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.937623 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.937635 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.937655 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.937670 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:57Z","lastTransitionTime":"2025-10-04T10:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:57 crc kubenswrapper[5025]: E1004 10:34:57.952096 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.961062 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.961096 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.961107 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.961123 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.961133 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:57Z","lastTransitionTime":"2025-10-04T10:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:57 crc kubenswrapper[5025]: E1004 10:34:57.983972 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:57Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:57 crc kubenswrapper[5025]: E1004 10:34:57.984178 5025 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.985812 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.985843 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.985855 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.985874 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:57 crc kubenswrapper[5025]: I1004 10:34:57.985888 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:57Z","lastTransitionTime":"2025-10-04T10:34:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.089797 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.090150 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.090250 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.090384 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.090538 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:58Z","lastTransitionTime":"2025-10-04T10:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.194666 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.194712 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.194758 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.194776 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.194788 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:58Z","lastTransitionTime":"2025-10-04T10:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.298045 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.298105 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.298119 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.298140 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.298152 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:58Z","lastTransitionTime":"2025-10-04T10:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.401900 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.402089 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.402122 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.402159 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.402189 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:58Z","lastTransitionTime":"2025-10-04T10:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.410589 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.410654 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.410599 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:34:58 crc kubenswrapper[5025]: E1004 10:34:58.410777 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:34:58 crc kubenswrapper[5025]: E1004 10:34:58.410986 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:34:58 crc kubenswrapper[5025]: E1004 10:34:58.411135 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.506181 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.506244 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.506265 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.506291 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.506312 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:58Z","lastTransitionTime":"2025-10-04T10:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.610956 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.611103 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.611126 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.611152 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.611173 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:58Z","lastTransitionTime":"2025-10-04T10:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.648367 5025 generic.go:334] "Generic (PLEG): container finished" podID="24e4becb-78de-4912-ac5f-d25ee8710f40" containerID="eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082" exitCode=0 Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.648494 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" event={"ID":"24e4becb-78de-4912-ac5f-d25ee8710f40","Type":"ContainerDied","Data":"eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.658891 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.677806 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.698003 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.714113 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.714199 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.714224 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.714258 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.714285 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:58Z","lastTransitionTime":"2025-10-04T10:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.718796 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.737939 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.771417 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.788668 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-wfm6t"] Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.789180 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.791092 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.791258 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.793381 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.793650 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.796823 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.811318 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs86s\" (UniqueName: \"kubernetes.io/projected/22480b6b-5db5-4368-b3f5-94305571cbff-kube-api-access-vs86s\") pod \"node-ca-wfm6t\" (UID: \"22480b6b-5db5-4368-b3f5-94305571cbff\") " pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.811384 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22480b6b-5db5-4368-b3f5-94305571cbff-host\") pod \"node-ca-wfm6t\" (UID: \"22480b6b-5db5-4368-b3f5-94305571cbff\") " pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.811421 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22480b6b-5db5-4368-b3f5-94305571cbff-serviceca\") pod \"node-ca-wfm6t\" (UID: \"22480b6b-5db5-4368-b3f5-94305571cbff\") " pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.811513 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.817519 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.817566 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.817582 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.817600 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.817611 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:58Z","lastTransitionTime":"2025-10-04T10:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.822891 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.838540 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.856458 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.873455 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.890238 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.910183 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.912580 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22480b6b-5db5-4368-b3f5-94305571cbff-host\") pod \"node-ca-wfm6t\" (UID: \"22480b6b-5db5-4368-b3f5-94305571cbff\") " pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.912622 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22480b6b-5db5-4368-b3f5-94305571cbff-serviceca\") pod \"node-ca-wfm6t\" (UID: \"22480b6b-5db5-4368-b3f5-94305571cbff\") " pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.912662 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs86s\" (UniqueName: \"kubernetes.io/projected/22480b6b-5db5-4368-b3f5-94305571cbff-kube-api-access-vs86s\") pod \"node-ca-wfm6t\" (UID: \"22480b6b-5db5-4368-b3f5-94305571cbff\") " pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.912704 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/22480b6b-5db5-4368-b3f5-94305571cbff-host\") pod \"node-ca-wfm6t\" (UID: \"22480b6b-5db5-4368-b3f5-94305571cbff\") " pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.913984 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/22480b6b-5db5-4368-b3f5-94305571cbff-serviceca\") pod \"node-ca-wfm6t\" (UID: \"22480b6b-5db5-4368-b3f5-94305571cbff\") " pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.920686 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.920742 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.920762 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.920787 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.920806 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:58Z","lastTransitionTime":"2025-10-04T10:34:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.926778 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.939099 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs86s\" (UniqueName: \"kubernetes.io/projected/22480b6b-5db5-4368-b3f5-94305571cbff-kube-api-access-vs86s\") pod \"node-ca-wfm6t\" (UID: \"22480b6b-5db5-4368-b3f5-94305571cbff\") " pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.943810 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.957410 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.971633 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:58 crc kubenswrapper[5025]: I1004 10:34:58.993463 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:58Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.013705 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.023661 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.023923 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.024078 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.024219 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.024347 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:59Z","lastTransitionTime":"2025-10-04T10:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.032214 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.044701 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.081041 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.103825 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.104203 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-wfm6t" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.123591 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: W1004 10:34:59.127151 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22480b6b_5db5_4368_b3f5_94305571cbff.slice/crio-820cfe445ecca4f80ffcda2845cbce56114577b8674f17a0ab80f770f6a960bf WatchSource:0}: Error finding container 820cfe445ecca4f80ffcda2845cbce56114577b8674f17a0ab80f770f6a960bf: Status 404 returned error can't find the container with id 820cfe445ecca4f80ffcda2845cbce56114577b8674f17a0ab80f770f6a960bf Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.128698 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.128743 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.128758 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.128780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.128797 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:59Z","lastTransitionTime":"2025-10-04T10:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.151767 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.176650 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.196131 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.208643 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.223416 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.231420 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.231467 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.231482 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.231505 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.231520 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:59Z","lastTransitionTime":"2025-10-04T10:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.335178 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.335240 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.335260 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.335288 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.335306 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:59Z","lastTransitionTime":"2025-10-04T10:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.438948 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.439098 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.439119 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.439154 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.439175 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:59Z","lastTransitionTime":"2025-10-04T10:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.543972 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.545220 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.545262 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.545303 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.545329 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:59Z","lastTransitionTime":"2025-10-04T10:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.651454 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.651492 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.651500 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.651516 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.651526 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:59Z","lastTransitionTime":"2025-10-04T10:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.669288 5025 generic.go:334] "Generic (PLEG): container finished" podID="24e4becb-78de-4912-ac5f-d25ee8710f40" containerID="6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837" exitCode=0 Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.669381 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" event={"ID":"24e4becb-78de-4912-ac5f-d25ee8710f40","Type":"ContainerDied","Data":"6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.673945 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wfm6t" event={"ID":"22480b6b-5db5-4368-b3f5-94305571cbff","Type":"ContainerStarted","Data":"820cfe445ecca4f80ffcda2845cbce56114577b8674f17a0ab80f770f6a960bf"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.694406 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.717279 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.732428 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.751242 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.756108 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.756182 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.756206 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.756237 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.756258 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:59Z","lastTransitionTime":"2025-10-04T10:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.768799 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.784897 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.802309 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.818300 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.835934 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.847317 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.858829 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.858874 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.858886 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.858904 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.858941 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:59Z","lastTransitionTime":"2025-10-04T10:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.875939 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.893201 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.910658 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.928839 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.942254 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:34:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.962950 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.962990 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.963002 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.963082 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:34:59 crc kubenswrapper[5025]: I1004 10:34:59.963094 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:34:59Z","lastTransitionTime":"2025-10-04T10:34:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.024861 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.025122 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.025167 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.025266 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:35:08.025189741 +0000 UTC m=+36.450156661 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.025346 5025 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.025391 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.025402 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.025419 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:08.025406507 +0000 UTC m=+36.450373397 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.025427 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.025569 5025 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.025637 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:08.025622233 +0000 UTC m=+36.450589154 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.025475 5025 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.025738 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:08.025712116 +0000 UTC m=+36.450679006 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.065641 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.065698 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.065718 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.065743 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.065761 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:00Z","lastTransitionTime":"2025-10-04T10:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.126269 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.126552 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.126601 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.126620 5025 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.126914 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:08.126879251 +0000 UTC m=+36.551846141 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.169627 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.169699 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.169722 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.169752 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.169771 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:00Z","lastTransitionTime":"2025-10-04T10:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.274308 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.274376 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.274396 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.274422 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.274440 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:00Z","lastTransitionTime":"2025-10-04T10:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.377338 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.377408 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.377430 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.377456 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.377475 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:00Z","lastTransitionTime":"2025-10-04T10:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.411361 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.411457 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.411366 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.411570 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.411689 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:00 crc kubenswrapper[5025]: E1004 10:35:00.411764 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.481754 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.481823 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.481854 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.481880 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.481899 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:00Z","lastTransitionTime":"2025-10-04T10:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.585523 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.585568 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.585581 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.585600 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.585610 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:00Z","lastTransitionTime":"2025-10-04T10:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.688727 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.689337 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.689355 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.689383 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.689400 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:00Z","lastTransitionTime":"2025-10-04T10:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.688968 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.690411 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.690531 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.701161 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" event={"ID":"24e4becb-78de-4912-ac5f-d25ee8710f40","Type":"ContainerStarted","Data":"27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.705342 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-wfm6t" event={"ID":"22480b6b-5db5-4368-b3f5-94305571cbff","Type":"ContainerStarted","Data":"a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.721775 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.729224 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.730757 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.743428 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.763455 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.780722 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.792213 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.792267 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.792282 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.792300 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.792354 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:00Z","lastTransitionTime":"2025-10-04T10:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.796064 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.810230 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.822306 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.841278 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.853631 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.865265 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.878231 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.890504 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.897492 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.897561 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.897581 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.897609 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.897644 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:00Z","lastTransitionTime":"2025-10-04T10:35:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.914475 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.928730 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.941543 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.957349 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.973696 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:00 crc kubenswrapper[5025]: I1004 10:35:00.991986 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:00Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.002597 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.002642 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.002656 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.002678 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.002697 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:01Z","lastTransitionTime":"2025-10-04T10:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.008874 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.024784 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.056834 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.073195 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.090487 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.104860 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.105559 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.105602 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.105613 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.105633 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.105643 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:01Z","lastTransitionTime":"2025-10-04T10:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.120567 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.134260 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.156032 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.176814 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.190423 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.204390 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:01Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.208357 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.208401 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.208417 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.208436 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.208449 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:01Z","lastTransitionTime":"2025-10-04T10:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.312111 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.312276 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.312294 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.312323 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.312343 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:01Z","lastTransitionTime":"2025-10-04T10:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.415462 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.415552 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.415571 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.415608 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.415627 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:01Z","lastTransitionTime":"2025-10-04T10:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.519389 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.519496 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.519531 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.519571 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.519601 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:01Z","lastTransitionTime":"2025-10-04T10:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.623727 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.623795 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.623814 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.623836 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.623851 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:01Z","lastTransitionTime":"2025-10-04T10:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.709925 5025 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.727455 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.727546 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.727564 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.727592 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.727612 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:01Z","lastTransitionTime":"2025-10-04T10:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.836513 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.836581 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.836591 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.836609 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.836620 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:01Z","lastTransitionTime":"2025-10-04T10:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.940042 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.940078 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.940088 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.940106 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:01 crc kubenswrapper[5025]: I1004 10:35:01.940118 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:01Z","lastTransitionTime":"2025-10-04T10:35:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.043044 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.043103 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.043118 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.043141 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.043157 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:02Z","lastTransitionTime":"2025-10-04T10:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.147199 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.147262 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.147277 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.147302 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.147318 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:02Z","lastTransitionTime":"2025-10-04T10:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.251180 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.251244 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.251261 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.251283 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.251299 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:02Z","lastTransitionTime":"2025-10-04T10:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.359805 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.359845 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.359853 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.359871 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.359883 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:02Z","lastTransitionTime":"2025-10-04T10:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.413208 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.413249 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:02 crc kubenswrapper[5025]: E1004 10:35:02.413420 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.413509 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:02 crc kubenswrapper[5025]: E1004 10:35:02.413662 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:02 crc kubenswrapper[5025]: E1004 10:35:02.413798 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.430413 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.453170 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.462956 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.463043 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.463059 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.463080 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.463116 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:02Z","lastTransitionTime":"2025-10-04T10:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.475535 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.503855 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.522702 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.542413 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.557375 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.566290 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.566352 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.566368 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.566391 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.566404 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:02Z","lastTransitionTime":"2025-10-04T10:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.570235 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.584700 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.610370 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.635548 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.654830 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.669258 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.669304 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.669314 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.669357 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.669367 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:02Z","lastTransitionTime":"2025-10-04T10:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.703094 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.712394 5025 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.728950 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.758553 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.772351 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.772386 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.772399 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.772429 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.772444 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:02Z","lastTransitionTime":"2025-10-04T10:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.874865 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.874936 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.874950 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.874971 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.874988 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:02Z","lastTransitionTime":"2025-10-04T10:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.978046 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.978115 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.978130 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.978148 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:02 crc kubenswrapper[5025]: I1004 10:35:02.978181 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:02Z","lastTransitionTime":"2025-10-04T10:35:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.081553 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.081633 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.081651 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.081681 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.081700 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:03Z","lastTransitionTime":"2025-10-04T10:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.184996 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.185475 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.185642 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.185799 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.186057 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:03Z","lastTransitionTime":"2025-10-04T10:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.289550 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.290096 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.290305 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.290514 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.290749 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:03Z","lastTransitionTime":"2025-10-04T10:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.394358 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.394438 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.394463 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.394494 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.394513 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:03Z","lastTransitionTime":"2025-10-04T10:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.499076 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.499393 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.499413 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.499435 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.499446 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:03Z","lastTransitionTime":"2025-10-04T10:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.606165 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.606215 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.606225 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.606241 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.606251 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:03Z","lastTransitionTime":"2025-10-04T10:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.709441 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.709529 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.709550 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.709570 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.709582 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:03Z","lastTransitionTime":"2025-10-04T10:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.813062 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.813620 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.813640 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.813678 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.813705 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:03Z","lastTransitionTime":"2025-10-04T10:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.919753 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.919823 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.919842 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.919869 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:03 crc kubenswrapper[5025]: I1004 10:35:03.919889 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:03Z","lastTransitionTime":"2025-10-04T10:35:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.024676 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.024722 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.024734 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.024752 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.024766 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:04Z","lastTransitionTime":"2025-10-04T10:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.128639 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.129444 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.129655 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.129799 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.129954 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:04Z","lastTransitionTime":"2025-10-04T10:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.233542 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.233942 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.234158 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.234325 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.234462 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:04Z","lastTransitionTime":"2025-10-04T10:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.338171 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.338749 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.339196 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.339445 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.339676 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:04Z","lastTransitionTime":"2025-10-04T10:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.410781 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:04 crc kubenswrapper[5025]: E1004 10:35:04.410972 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.411630 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:04 crc kubenswrapper[5025]: E1004 10:35:04.411742 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.411825 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:04 crc kubenswrapper[5025]: E1004 10:35:04.411904 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.449190 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.449311 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.449330 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.449356 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.449374 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:04Z","lastTransitionTime":"2025-10-04T10:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.552063 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.552116 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.552126 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.552142 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.552155 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:04Z","lastTransitionTime":"2025-10-04T10:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.655503 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.655604 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.655668 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.655704 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.655732 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:04Z","lastTransitionTime":"2025-10-04T10:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.722561 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/0.log" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.727365 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5" exitCode=1 Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.727488 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.729306 5025 scope.go:117] "RemoveContainer" containerID="877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.760042 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.760505 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.760831 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.761102 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.761322 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:04Z","lastTransitionTime":"2025-10-04T10:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.779945 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.798571 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.814815 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.838085 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.861003 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.863914 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.863961 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.863974 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.863995 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.864026 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:04Z","lastTransitionTime":"2025-10-04T10:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.892780 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.916114 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.954917 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.966935 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.966987 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.966999 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.967034 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.967045 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:04Z","lastTransitionTime":"2025-10-04T10:35:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.974522 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:04 crc kubenswrapper[5025]: I1004 10:35:04.995145 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:04Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.018312 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:05Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.034702 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:05Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.055295 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:05Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.069627 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.069906 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.070306 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.071160 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.071220 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:05Z","lastTransitionTime":"2025-10-04T10:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.072577 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:05Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.097840 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:03Z\\\",\\\"message\\\":\\\"ons/factory.go:140\\\\nI1004 10:35:03.572960 6323 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.573074 6323 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573368 6323 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573657 6323 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573847 6323 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.574083 6323 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.574221 6323 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:35:03.574386 6323 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:05Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.175963 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.176046 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.176059 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.176082 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.176095 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:05Z","lastTransitionTime":"2025-10-04T10:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.279003 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.279086 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.279099 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.279121 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.279135 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:05Z","lastTransitionTime":"2025-10-04T10:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.381992 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.382110 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.382136 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.382165 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.382184 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:05Z","lastTransitionTime":"2025-10-04T10:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.485857 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.485912 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.485924 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.485945 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.485958 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:05Z","lastTransitionTime":"2025-10-04T10:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.589158 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.589258 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.589290 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.589325 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.589353 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:05Z","lastTransitionTime":"2025-10-04T10:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.692709 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.692812 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.692839 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.692882 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.692908 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:05Z","lastTransitionTime":"2025-10-04T10:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.734782 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/0.log" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.738542 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df"} Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.795478 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.795562 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.795582 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.795610 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.795632 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:05Z","lastTransitionTime":"2025-10-04T10:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.899777 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.899868 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.899893 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.899930 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:05 crc kubenswrapper[5025]: I1004 10:35:05.899954 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:05Z","lastTransitionTime":"2025-10-04T10:35:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.003243 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.003315 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.003330 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.003353 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.003370 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:06Z","lastTransitionTime":"2025-10-04T10:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.075383 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk"] Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.076837 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.079540 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.079555 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.095003 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.106846 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.106885 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.106894 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.106909 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.106920 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:06Z","lastTransitionTime":"2025-10-04T10:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.110106 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.129617 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.142281 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.167700 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:03Z\\\",\\\"message\\\":\\\"ons/factory.go:140\\\\nI1004 10:35:03.572960 6323 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.573074 6323 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573368 6323 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573657 6323 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573847 6323 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.574083 6323 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.574221 6323 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:35:03.574386 6323 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.186519 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.202659 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.209909 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.210055 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.210127 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.210199 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.210270 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:06Z","lastTransitionTime":"2025-10-04T10:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.213585 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdpgt\" (UniqueName: \"kubernetes.io/projected/44d57967-21de-4a17-9fe3-b5a7678e7a82-kube-api-access-bdpgt\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.213687 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/44d57967-21de-4a17-9fe3-b5a7678e7a82-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.213740 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/44d57967-21de-4a17-9fe3-b5a7678e7a82-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.213799 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/44d57967-21de-4a17-9fe3-b5a7678e7a82-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.243905 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.257352 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.291186 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.309112 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.313073 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.313250 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.313333 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.313427 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.313524 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:06Z","lastTransitionTime":"2025-10-04T10:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.314262 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/44d57967-21de-4a17-9fe3-b5a7678e7a82-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.314386 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdpgt\" (UniqueName: \"kubernetes.io/projected/44d57967-21de-4a17-9fe3-b5a7678e7a82-kube-api-access-bdpgt\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.314501 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/44d57967-21de-4a17-9fe3-b5a7678e7a82-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.314582 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/44d57967-21de-4a17-9fe3-b5a7678e7a82-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.315413 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/44d57967-21de-4a17-9fe3-b5a7678e7a82-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.315490 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/44d57967-21de-4a17-9fe3-b5a7678e7a82-env-overrides\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.327080 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/44d57967-21de-4a17-9fe3-b5a7678e7a82-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.334914 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.345154 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdpgt\" (UniqueName: \"kubernetes.io/projected/44d57967-21de-4a17-9fe3-b5a7678e7a82-kube-api-access-bdpgt\") pod \"ovnkube-control-plane-749d76644c-7hlrk\" (UID: \"44d57967-21de-4a17-9fe3-b5a7678e7a82\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.351630 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.367471 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.384992 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.391870 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.402228 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.410482 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:06 crc kubenswrapper[5025]: E1004 10:35:06.410701 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.410550 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:06 crc kubenswrapper[5025]: E1004 10:35:06.410888 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.410525 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:06 crc kubenswrapper[5025]: E1004 10:35:06.411091 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:06 crc kubenswrapper[5025]: W1004 10:35:06.413478 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44d57967_21de_4a17_9fe3_b5a7678e7a82.slice/crio-369f9b907d04ba75d7050b93597fc5596d83940a2c508cb89f1c08fdbb9f84ff WatchSource:0}: Error finding container 369f9b907d04ba75d7050b93597fc5596d83940a2c508cb89f1c08fdbb9f84ff: Status 404 returned error can't find the container with id 369f9b907d04ba75d7050b93597fc5596d83940a2c508cb89f1c08fdbb9f84ff Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.416280 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.416306 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.416317 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.416333 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.416342 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:06Z","lastTransitionTime":"2025-10-04T10:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.521694 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.521745 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.521759 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.521777 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.521793 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:06Z","lastTransitionTime":"2025-10-04T10:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.623821 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.623879 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.623893 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.623914 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.623929 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:06Z","lastTransitionTime":"2025-10-04T10:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.728230 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.728297 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.728318 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.728349 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.728370 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:06Z","lastTransitionTime":"2025-10-04T10:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.746028 5025 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.746081 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" event={"ID":"44d57967-21de-4a17-9fe3-b5a7678e7a82","Type":"ContainerStarted","Data":"e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.746172 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" event={"ID":"44d57967-21de-4a17-9fe3-b5a7678e7a82","Type":"ContainerStarted","Data":"369f9b907d04ba75d7050b93597fc5596d83940a2c508cb89f1c08fdbb9f84ff"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.774087 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.790694 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.802699 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.825135 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.831958 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.831999 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.832028 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.832047 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.832064 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:06Z","lastTransitionTime":"2025-10-04T10:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.841328 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.858231 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.874152 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.889791 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.910180 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.923838 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.934935 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.934992 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.935005 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.935062 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.935074 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:06Z","lastTransitionTime":"2025-10-04T10:35:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.940693 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.959531 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.979568 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:06 crc kubenswrapper[5025]: I1004 10:35:06.992539 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:06Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.009341 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.029345 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:03Z\\\",\\\"message\\\":\\\"ons/factory.go:140\\\\nI1004 10:35:03.572960 6323 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.573074 6323 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573368 6323 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573657 6323 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573847 6323 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.574083 6323 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.574221 6323 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:35:03.574386 6323 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.038752 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.038815 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.038832 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.038861 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.038875 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:07Z","lastTransitionTime":"2025-10-04T10:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.142157 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.142218 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.142232 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.142251 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.142262 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:07Z","lastTransitionTime":"2025-10-04T10:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.245485 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.245521 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.245531 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.245547 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.245558 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:07Z","lastTransitionTime":"2025-10-04T10:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.347860 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.347911 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.347923 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.347944 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.347957 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:07Z","lastTransitionTime":"2025-10-04T10:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.451181 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.451240 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.451252 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.451273 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.451284 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:07Z","lastTransitionTime":"2025-10-04T10:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.537097 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-frc27"] Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.537927 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:07 crc kubenswrapper[5025]: E1004 10:35:07.538084 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.553699 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.553738 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.553749 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.553764 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.553773 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:07Z","lastTransitionTime":"2025-10-04T10:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.555416 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.566273 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.579430 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.591758 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.605835 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.619639 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.630214 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.630261 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5hvf\" (UniqueName: \"kubernetes.io/projected/86312d38-63d0-409a-98d4-727f0fb47929-kube-api-access-x5hvf\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.631517 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.641414 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.656655 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.656704 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.656715 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.656734 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.656749 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:07Z","lastTransitionTime":"2025-10-04T10:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.660621 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:03Z\\\",\\\"message\\\":\\\"ons/factory.go:140\\\\nI1004 10:35:03.572960 6323 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.573074 6323 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573368 6323 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573657 6323 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573847 6323 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.574083 6323 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.574221 6323 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:35:03.574386 6323 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.672457 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.686803 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.701181 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.714591 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.727935 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.731296 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5hvf\" (UniqueName: \"kubernetes.io/projected/86312d38-63d0-409a-98d4-727f0fb47929-kube-api-access-x5hvf\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.731362 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:07 crc kubenswrapper[5025]: E1004 10:35:07.731480 5025 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:07 crc kubenswrapper[5025]: E1004 10:35:07.731524 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs podName:86312d38-63d0-409a-98d4-727f0fb47929 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:08.231507782 +0000 UTC m=+36.656474662 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs") pod "network-metrics-daemon-frc27" (UID: "86312d38-63d0-409a-98d4-727f0fb47929") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.747601 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5hvf\" (UniqueName: \"kubernetes.io/projected/86312d38-63d0-409a-98d4-727f0fb47929-kube-api-access-x5hvf\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.749006 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.757681 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/1.log" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.758318 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/0.log" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.758446 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.758487 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.758499 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.758517 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.758529 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:07Z","lastTransitionTime":"2025-10-04T10:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.761188 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df" exitCode=1 Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.761262 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.761465 5025 scope.go:117] "RemoveContainer" containerID="877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.762224 5025 scope.go:117] "RemoveContainer" containerID="c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df" Oct 04 10:35:07 crc kubenswrapper[5025]: E1004 10:35:07.762426 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.763549 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" event={"ID":"44d57967-21de-4a17-9fe3-b5a7678e7a82","Type":"ContainerStarted","Data":"68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.764947 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.781992 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.797488 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.810140 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.823282 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.840415 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.861245 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:03Z\\\",\\\"message\\\":\\\"ons/factory.go:140\\\\nI1004 10:35:03.572960 6323 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.573074 6323 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573368 6323 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573657 6323 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573847 6323 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.574083 6323 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.574221 6323 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:35:03.574386 6323 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"il,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1004 10:35:06.985348 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:06.985359 6462 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-operator for network=default are: map[]\\\\nI1004 10:35:06.985373 6462 services_controller.go:443] Built service openshift-machine-api/machine-api-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.21\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 10:35:06.985386 6462 services_controller.go:444] Built service openshift-machine-api/machine-api-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 10:35:06.985394 6462 services_controller.go:445] Built service openshift-machine-api/machine-api-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nF1004 10:35:06.985431 6462 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.861893 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.861922 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.861934 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.861952 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.861964 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:07Z","lastTransitionTime":"2025-10-04T10:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.872266 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.890176 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.902743 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.915886 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.931435 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.946885 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.964870 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.965421 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.965504 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.965528 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.965560 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.965581 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:07Z","lastTransitionTime":"2025-10-04T10:35:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.978790 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:07 crc kubenswrapper[5025]: I1004 10:35:07.996845 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:07Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.009654 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:08Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.026617 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:08Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.034974 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.035110 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.035140 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.035164 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:35:24.035135942 +0000 UTC m=+52.460102822 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.035219 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.035253 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.035269 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.035281 5025 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.035301 5025 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.035323 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:24.035308327 +0000 UTC m=+52.460275207 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.035356 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:24.035345318 +0000 UTC m=+52.460312188 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.035372 5025 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.035400 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:24.035393129 +0000 UTC m=+52.460360009 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.044343 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:08Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.068696 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.068728 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.068739 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.068754 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.068764 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.121663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.121715 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.121729 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.121745 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.121757 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.135175 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:08Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.136153 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.136367 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.136403 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.136419 5025 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.136496 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:24.136472102 +0000 UTC m=+52.561439052 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.139731 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.139768 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.139780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.139797 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.139810 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.153155 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:08Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.157181 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.157222 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.157230 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.157250 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.157260 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.172911 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:08Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.177821 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.177865 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.177897 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.177917 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.177932 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.194879 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:08Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.198800 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.198850 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.198863 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.198881 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.198893 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.212770 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:08Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.212906 5025 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.214654 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.214716 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.214728 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.214748 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.214761 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.237310 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.237469 5025 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.237530 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs podName:86312d38-63d0-409a-98d4-727f0fb47929 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:09.237510583 +0000 UTC m=+37.662477473 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs") pod "network-metrics-daemon-frc27" (UID: "86312d38-63d0-409a-98d4-727f0fb47929") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.318126 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.318175 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.318187 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.318204 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.318218 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.411399 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.411401 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.411569 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.411698 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.411405 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:08 crc kubenswrapper[5025]: E1004 10:35:08.411973 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.420209 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.420255 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.420267 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.420287 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.420300 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.522829 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.522928 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.522942 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.522958 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.522989 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.626422 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.626476 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.626488 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.626504 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.626514 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.729405 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.729451 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.729462 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.729478 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.729488 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.769126 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/1.log" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.832048 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.832093 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.832105 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.832121 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.832134 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.935105 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.935150 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.935160 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.935180 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:08 crc kubenswrapper[5025]: I1004 10:35:08.935193 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:08Z","lastTransitionTime":"2025-10-04T10:35:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.038791 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.038842 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.038856 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.038872 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.038886 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:09Z","lastTransitionTime":"2025-10-04T10:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.142226 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.142278 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.142290 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.142307 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.142318 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:09Z","lastTransitionTime":"2025-10-04T10:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.245065 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.245108 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.245118 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.245132 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.245143 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:09Z","lastTransitionTime":"2025-10-04T10:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.248736 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:09 crc kubenswrapper[5025]: E1004 10:35:09.248861 5025 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:09 crc kubenswrapper[5025]: E1004 10:35:09.248920 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs podName:86312d38-63d0-409a-98d4-727f0fb47929 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:11.248901916 +0000 UTC m=+39.673868796 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs") pod "network-metrics-daemon-frc27" (UID: "86312d38-63d0-409a-98d4-727f0fb47929") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.348975 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.349146 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.349163 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.349185 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.349207 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:09Z","lastTransitionTime":"2025-10-04T10:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.411235 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:09 crc kubenswrapper[5025]: E1004 10:35:09.411505 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.452754 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.452792 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.452800 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.452813 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.452822 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:09Z","lastTransitionTime":"2025-10-04T10:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.556290 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.556341 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.556358 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.556383 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.556400 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:09Z","lastTransitionTime":"2025-10-04T10:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.660592 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.660658 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.660670 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.660699 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.660714 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:09Z","lastTransitionTime":"2025-10-04T10:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.764433 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.764662 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.764672 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.764687 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.764698 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:09Z","lastTransitionTime":"2025-10-04T10:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.867334 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.867374 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.867384 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.867399 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.867408 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:09Z","lastTransitionTime":"2025-10-04T10:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.970428 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.970494 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.970508 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.970536 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:09 crc kubenswrapper[5025]: I1004 10:35:09.970548 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:09Z","lastTransitionTime":"2025-10-04T10:35:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.073969 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.074475 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.074645 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.074832 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.075009 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:10Z","lastTransitionTime":"2025-10-04T10:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.178835 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.179490 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.179551 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.179575 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.179590 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:10Z","lastTransitionTime":"2025-10-04T10:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.283252 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.283304 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.283314 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.283331 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.283343 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:10Z","lastTransitionTime":"2025-10-04T10:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.386269 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.386328 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.386339 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.386357 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.386368 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:10Z","lastTransitionTime":"2025-10-04T10:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.410276 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.410310 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.410352 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:10 crc kubenswrapper[5025]: E1004 10:35:10.410557 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:10 crc kubenswrapper[5025]: E1004 10:35:10.410676 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:10 crc kubenswrapper[5025]: E1004 10:35:10.410823 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.489833 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.489928 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.489960 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.489996 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.490051 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:10Z","lastTransitionTime":"2025-10-04T10:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.593697 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.593780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.593808 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.593841 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.593864 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:10Z","lastTransitionTime":"2025-10-04T10:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.697447 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.697507 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.697519 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.697541 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.697553 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:10Z","lastTransitionTime":"2025-10-04T10:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.803510 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.803606 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.803629 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.803663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.803718 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:10Z","lastTransitionTime":"2025-10-04T10:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.907351 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.907408 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.907422 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.907441 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:10 crc kubenswrapper[5025]: I1004 10:35:10.907455 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:10Z","lastTransitionTime":"2025-10-04T10:35:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.011217 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.012257 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.012311 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.012338 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.012359 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:11Z","lastTransitionTime":"2025-10-04T10:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.115488 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.115561 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.115578 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.115598 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.115611 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:11Z","lastTransitionTime":"2025-10-04T10:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.219148 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.219218 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.219230 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.219252 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.219268 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:11Z","lastTransitionTime":"2025-10-04T10:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.284231 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:11 crc kubenswrapper[5025]: E1004 10:35:11.284457 5025 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:11 crc kubenswrapper[5025]: E1004 10:35:11.284597 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs podName:86312d38-63d0-409a-98d4-727f0fb47929 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:15.284555904 +0000 UTC m=+43.709522824 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs") pod "network-metrics-daemon-frc27" (UID: "86312d38-63d0-409a-98d4-727f0fb47929") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.322940 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.323004 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.323039 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.323062 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.323075 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:11Z","lastTransitionTime":"2025-10-04T10:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.410539 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:11 crc kubenswrapper[5025]: E1004 10:35:11.410867 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.426003 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.426131 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.426156 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.426188 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.426212 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:11Z","lastTransitionTime":"2025-10-04T10:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.530230 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.530308 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.530348 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.530372 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.530385 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:11Z","lastTransitionTime":"2025-10-04T10:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.632688 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.632747 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.632759 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.632777 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.632789 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:11Z","lastTransitionTime":"2025-10-04T10:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.735739 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.735795 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.735807 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.735824 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.735836 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:11Z","lastTransitionTime":"2025-10-04T10:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.839905 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.839959 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.839975 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.840000 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.840046 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:11Z","lastTransitionTime":"2025-10-04T10:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.943359 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.943406 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.943414 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.943432 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.943443 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:11Z","lastTransitionTime":"2025-10-04T10:35:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.970229 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.982716 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:11Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:11 crc kubenswrapper[5025]: I1004 10:35:11.996441 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:11Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.008787 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.019465 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.036711 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:03Z\\\",\\\"message\\\":\\\"ons/factory.go:140\\\\nI1004 10:35:03.572960 6323 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.573074 6323 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573368 6323 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573657 6323 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573847 6323 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.574083 6323 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.574221 6323 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:35:03.574386 6323 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"il,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1004 10:35:06.985348 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:06.985359 6462 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-operator for network=default are: map[]\\\\nI1004 10:35:06.985373 6462 services_controller.go:443] Built service openshift-machine-api/machine-api-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.21\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 10:35:06.985386 6462 services_controller.go:444] Built service openshift-machine-api/machine-api-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 10:35:06.985394 6462 services_controller.go:445] Built service openshift-machine-api/machine-api-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nF1004 10:35:06.985431 6462 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.047035 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.047325 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.047413 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.047501 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.047580 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:12Z","lastTransitionTime":"2025-10-04T10:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.048186 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.061359 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.078169 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.103979 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.116691 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.136183 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.150335 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.150713 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.150795 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.150888 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.150974 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:12Z","lastTransitionTime":"2025-10-04T10:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.153049 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.166396 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.181599 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.195870 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.212814 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.227425 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.254137 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.254190 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.254201 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.254220 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.254232 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:12Z","lastTransitionTime":"2025-10-04T10:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.357881 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.358280 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.358346 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.358470 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.358540 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:12Z","lastTransitionTime":"2025-10-04T10:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.410898 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.410989 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:12 crc kubenswrapper[5025]: E1004 10:35:12.411064 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:12 crc kubenswrapper[5025]: E1004 10:35:12.411213 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.411471 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:12 crc kubenswrapper[5025]: E1004 10:35:12.411695 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.423720 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.436896 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.451031 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.461269 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.461304 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.461315 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.461330 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.461341 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:12Z","lastTransitionTime":"2025-10-04T10:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.464096 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.476478 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.507521 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://877673169a2b040dd2987a3775ea0f255c04d258ba76c54c4108c1598b52faa5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:03Z\\\",\\\"message\\\":\\\"ons/factory.go:140\\\\nI1004 10:35:03.572960 6323 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.573074 6323 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573368 6323 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573657 6323 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.573847 6323 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1004 10:35:03.574083 6323 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1004 10:35:03.574221 6323 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1004 10:35:03.574386 6323 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"il,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1004 10:35:06.985348 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:06.985359 6462 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-operator for network=default are: map[]\\\\nI1004 10:35:06.985373 6462 services_controller.go:443] Built service openshift-machine-api/machine-api-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.21\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 10:35:06.985386 6462 services_controller.go:444] Built service openshift-machine-api/machine-api-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 10:35:06.985394 6462 services_controller.go:445] Built service openshift-machine-api/machine-api-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nF1004 10:35:06.985431 6462 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.523792 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.541390 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.554531 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.564289 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.564350 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.564362 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.564385 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.564398 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:12Z","lastTransitionTime":"2025-10-04T10:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.568129 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.586186 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.598247 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.607439 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.619488 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.633144 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.645317 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.659479 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:12Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.667202 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.667300 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.667327 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.667364 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.667392 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:12Z","lastTransitionTime":"2025-10-04T10:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.771121 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.771172 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.771184 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.771204 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.771220 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:12Z","lastTransitionTime":"2025-10-04T10:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.874974 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.875030 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.875040 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.875056 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.875066 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:12Z","lastTransitionTime":"2025-10-04T10:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.978493 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.978546 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.978556 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.978575 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:12 crc kubenswrapper[5025]: I1004 10:35:12.978587 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:12Z","lastTransitionTime":"2025-10-04T10:35:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.082263 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.082335 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.082354 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.082381 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.082400 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:13Z","lastTransitionTime":"2025-10-04T10:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.185411 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.185476 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.185495 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.185520 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.185538 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:13Z","lastTransitionTime":"2025-10-04T10:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.289105 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.289191 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.289216 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.289252 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.289275 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:13Z","lastTransitionTime":"2025-10-04T10:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.392377 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.392449 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.392474 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.392507 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.392531 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:13Z","lastTransitionTime":"2025-10-04T10:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.410345 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:13 crc kubenswrapper[5025]: E1004 10:35:13.410584 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.495920 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.496358 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.496373 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.496393 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.496408 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:13Z","lastTransitionTime":"2025-10-04T10:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.599928 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.599994 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.600028 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.600046 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.600057 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:13Z","lastTransitionTime":"2025-10-04T10:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.702847 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.702894 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.702903 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.702921 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.702932 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:13Z","lastTransitionTime":"2025-10-04T10:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.805504 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.805550 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.805560 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.805578 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.805591 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:13Z","lastTransitionTime":"2025-10-04T10:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.908049 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.908109 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.908142 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.908170 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:13 crc kubenswrapper[5025]: I1004 10:35:13.908188 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:13Z","lastTransitionTime":"2025-10-04T10:35:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.011660 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.011747 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.011757 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.011794 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.011807 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:14Z","lastTransitionTime":"2025-10-04T10:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.114780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.114826 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.114837 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.114851 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.114861 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:14Z","lastTransitionTime":"2025-10-04T10:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.217726 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.218138 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.218326 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.218442 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.218548 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:14Z","lastTransitionTime":"2025-10-04T10:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.321762 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.321815 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.321829 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.321847 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.321861 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:14Z","lastTransitionTime":"2025-10-04T10:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.410496 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.410583 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.410706 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:14 crc kubenswrapper[5025]: E1004 10:35:14.410711 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:14 crc kubenswrapper[5025]: E1004 10:35:14.410831 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:14 crc kubenswrapper[5025]: E1004 10:35:14.410944 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.424108 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.424158 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.424198 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.424224 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.424235 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:14Z","lastTransitionTime":"2025-10-04T10:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.527307 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.527350 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.527361 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.527376 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.527390 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:14Z","lastTransitionTime":"2025-10-04T10:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.629783 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.629839 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.629851 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.629876 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.629891 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:14Z","lastTransitionTime":"2025-10-04T10:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.732722 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.732774 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.732786 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.732804 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.732815 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:14Z","lastTransitionTime":"2025-10-04T10:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.835621 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.835714 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.835734 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.835791 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.835811 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:14Z","lastTransitionTime":"2025-10-04T10:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.938521 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.938588 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.938610 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.938639 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:14 crc kubenswrapper[5025]: I1004 10:35:14.938664 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:14Z","lastTransitionTime":"2025-10-04T10:35:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.041751 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.041809 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.041824 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.041849 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.041863 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:15Z","lastTransitionTime":"2025-10-04T10:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.144709 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.144747 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.144757 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.144772 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.144782 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:15Z","lastTransitionTime":"2025-10-04T10:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.247072 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.247113 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.247122 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.247139 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.247150 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:15Z","lastTransitionTime":"2025-10-04T10:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.326150 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:15 crc kubenswrapper[5025]: E1004 10:35:15.326326 5025 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:15 crc kubenswrapper[5025]: E1004 10:35:15.326390 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs podName:86312d38-63d0-409a-98d4-727f0fb47929 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:23.326372603 +0000 UTC m=+51.751339483 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs") pod "network-metrics-daemon-frc27" (UID: "86312d38-63d0-409a-98d4-727f0fb47929") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.349691 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.349729 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.349738 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.349752 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.349762 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:15Z","lastTransitionTime":"2025-10-04T10:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.410710 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:15 crc kubenswrapper[5025]: E1004 10:35:15.410962 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.452486 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.452544 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.452559 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.452584 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.452599 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:15Z","lastTransitionTime":"2025-10-04T10:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.556789 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.556840 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.556850 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.556867 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.556879 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:15Z","lastTransitionTime":"2025-10-04T10:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.659139 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.659179 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.659189 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.659228 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.659246 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:15Z","lastTransitionTime":"2025-10-04T10:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.762530 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.762590 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.762607 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.762634 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.762653 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:15Z","lastTransitionTime":"2025-10-04T10:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.865773 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.865834 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.865848 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.865868 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.865884 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:15Z","lastTransitionTime":"2025-10-04T10:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.969244 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.969683 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.969883 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.970175 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:15 crc kubenswrapper[5025]: I1004 10:35:15.970376 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:15Z","lastTransitionTime":"2025-10-04T10:35:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.073577 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.073955 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.074111 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.074208 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.074293 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:16Z","lastTransitionTime":"2025-10-04T10:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.177345 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.177409 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.177426 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.177446 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.177456 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:16Z","lastTransitionTime":"2025-10-04T10:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.281304 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.281627 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.281691 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.281760 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.281883 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:16Z","lastTransitionTime":"2025-10-04T10:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.385512 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.385835 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.385952 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.386065 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.386130 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:16Z","lastTransitionTime":"2025-10-04T10:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.411094 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.411315 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.411235 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:16 crc kubenswrapper[5025]: E1004 10:35:16.411563 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:16 crc kubenswrapper[5025]: E1004 10:35:16.411679 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:16 crc kubenswrapper[5025]: E1004 10:35:16.411898 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.489104 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.489193 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.489219 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.489251 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.489277 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:16Z","lastTransitionTime":"2025-10-04T10:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.592680 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.592762 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.592784 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.592809 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.592825 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:16Z","lastTransitionTime":"2025-10-04T10:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.695442 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.695495 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.695508 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.695524 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.695539 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:16Z","lastTransitionTime":"2025-10-04T10:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.798737 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.799181 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.799314 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.799441 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.799553 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:16Z","lastTransitionTime":"2025-10-04T10:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.902677 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.902743 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.902754 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.902775 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:16 crc kubenswrapper[5025]: I1004 10:35:16.902785 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:16Z","lastTransitionTime":"2025-10-04T10:35:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.005612 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.005659 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.005674 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.005692 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.005707 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:17Z","lastTransitionTime":"2025-10-04T10:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.108531 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.108913 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.109002 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.109120 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.109233 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:17Z","lastTransitionTime":"2025-10-04T10:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.211684 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.212321 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.212392 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.212457 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.212565 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:17Z","lastTransitionTime":"2025-10-04T10:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.315164 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.315220 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.315235 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.315256 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.315273 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:17Z","lastTransitionTime":"2025-10-04T10:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.411091 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:17 crc kubenswrapper[5025]: E1004 10:35:17.411236 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.417710 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.417731 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.417739 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.417753 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.417761 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:17Z","lastTransitionTime":"2025-10-04T10:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.520797 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.520841 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.520849 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.520864 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.520874 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:17Z","lastTransitionTime":"2025-10-04T10:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.624124 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.624174 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.624196 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.624217 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.624232 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:17Z","lastTransitionTime":"2025-10-04T10:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.728305 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.728346 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.728356 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.728375 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.728400 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:17Z","lastTransitionTime":"2025-10-04T10:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.831259 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.831308 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.831320 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.831336 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.831348 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:17Z","lastTransitionTime":"2025-10-04T10:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.933957 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.934004 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.934035 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.934053 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:17 crc kubenswrapper[5025]: I1004 10:35:17.934063 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:17Z","lastTransitionTime":"2025-10-04T10:35:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.037513 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.037562 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.037573 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.037590 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.037601 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.140192 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.140255 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.140268 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.140290 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.140311 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.248810 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.248932 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.248954 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.249049 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.249076 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.352653 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.352712 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.352726 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.352752 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.352765 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.394665 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.394718 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.394729 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.394746 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.394759 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: E1004 10:35:18.409597 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.410630 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.410782 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:18 crc kubenswrapper[5025]: E1004 10:35:18.410914 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.410947 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:18 crc kubenswrapper[5025]: E1004 10:35:18.411361 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.411655 5025 scope.go:117] "RemoveContainer" containerID="c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df" Oct 04 10:35:18 crc kubenswrapper[5025]: E1004 10:35:18.411867 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.419667 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.419713 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.419726 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.419743 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.419754 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.439879 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"il,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1004 10:35:06.985348 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:06.985359 6462 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-operator for network=default are: map[]\\\\nI1004 10:35:06.985373 6462 services_controller.go:443] Built service openshift-machine-api/machine-api-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.21\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 10:35:06.985386 6462 services_controller.go:444] Built service openshift-machine-api/machine-api-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 10:35:06.985394 6462 services_controller.go:445] Built service openshift-machine-api/machine-api-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nF1004 10:35:06.985431 6462 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: E1004 10:35:18.443749 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.449709 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.449771 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.449785 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.449814 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.449834 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.455455 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: E1004 10:35:18.464516 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.469850 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.469899 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.469916 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.469931 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.469942 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.473346 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: E1004 10:35:18.484169 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.488718 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.488757 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.488769 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.488788 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.488800 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.490084 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: E1004 10:35:18.502605 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: E1004 10:35:18.502722 5025 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.504741 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.504775 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.504787 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.504805 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.504818 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.511681 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.539547 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.567973 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.613589 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.613645 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.613656 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.613676 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.613688 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.628652 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.642244 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.659608 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.677395 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.696281 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.710570 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.725327 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.725383 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.725396 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.725417 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.725431 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.725592 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.740811 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.755826 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.770075 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.824947 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/1.log" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.828211 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.828272 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.828287 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.828369 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.828386 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.829851 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.830104 5025 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.861986 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"il,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1004 10:35:06.985348 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:06.985359 6462 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-operator for network=default are: map[]\\\\nI1004 10:35:06.985373 6462 services_controller.go:443] Built service openshift-machine-api/machine-api-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.21\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 10:35:06.985386 6462 services_controller.go:444] Built service openshift-machine-api/machine-api-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 10:35:06.985394 6462 services_controller.go:445] Built service openshift-machine-api/machine-api-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nF1004 10:35:06.985431 6462 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.882759 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.904179 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.931470 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.931559 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.931587 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.931662 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.931693 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:18Z","lastTransitionTime":"2025-10-04T10:35:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.936504 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.962771 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.975586 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:18 crc kubenswrapper[5025]: I1004 10:35:18.995761 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:18Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.009402 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.024072 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.034808 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.034863 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.034876 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.034896 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.034910 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:19Z","lastTransitionTime":"2025-10-04T10:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.039739 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.054809 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.074712 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.089682 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.104698 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.119968 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.133406 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.137232 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.137294 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.137308 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.137328 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.137341 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:19Z","lastTransitionTime":"2025-10-04T10:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.150569 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.240214 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.240315 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.240333 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.240353 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.240369 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:19Z","lastTransitionTime":"2025-10-04T10:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.343668 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.343724 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.343738 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.343758 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.343770 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:19Z","lastTransitionTime":"2025-10-04T10:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.410422 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:19 crc kubenswrapper[5025]: E1004 10:35:19.410598 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.446369 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.446419 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.446430 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.446449 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.446462 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:19Z","lastTransitionTime":"2025-10-04T10:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.549371 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.549433 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.549822 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.549869 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.549883 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:19Z","lastTransitionTime":"2025-10-04T10:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.653793 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.653845 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.653854 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.653869 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.653880 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:19Z","lastTransitionTime":"2025-10-04T10:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.757740 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.757801 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.757813 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.757834 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.757848 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:19Z","lastTransitionTime":"2025-10-04T10:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.835407 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/2.log" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.836392 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/1.log" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.840273 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc" exitCode=1 Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.840336 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.840387 5025 scope.go:117] "RemoveContainer" containerID="c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.841132 5025 scope.go:117] "RemoveContainer" containerID="161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc" Oct 04 10:35:19 crc kubenswrapper[5025]: E1004 10:35:19.841316 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.862419 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.862479 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.862498 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.862526 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.862849 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:19Z","lastTransitionTime":"2025-10-04T10:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.871001 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.893678 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.916398 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.929058 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.957270 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"il,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1004 10:35:06.985348 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:06.985359 6462 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-operator for network=default are: map[]\\\\nI1004 10:35:06.985373 6462 services_controller.go:443] Built service openshift-machine-api/machine-api-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.21\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 10:35:06.985386 6462 services_controller.go:444] Built service openshift-machine-api/machine-api-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 10:35:06.985394 6462 services_controller.go:445] Built service openshift-machine-api/machine-api-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nF1004 10:35:06.985431 6462 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.965934 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.965982 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.965992 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.966015 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.966046 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:19Z","lastTransitionTime":"2025-10-04T10:35:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:19 crc kubenswrapper[5025]: I1004 10:35:19.973127 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:19Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.003486 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.016928 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.031382 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.050887 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.068786 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.069733 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.069766 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.069778 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.069798 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.069811 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:20Z","lastTransitionTime":"2025-10-04T10:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.084403 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.096656 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.111172 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.126500 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.139732 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.151126 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:20Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.172856 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.173060 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.173095 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.173127 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.173155 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:20Z","lastTransitionTime":"2025-10-04T10:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.276216 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.276356 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.276375 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.276403 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.276420 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:20Z","lastTransitionTime":"2025-10-04T10:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.379379 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.379429 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.379440 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.379459 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.379469 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:20Z","lastTransitionTime":"2025-10-04T10:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.410603 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.410630 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:20 crc kubenswrapper[5025]: E1004 10:35:20.410927 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.411098 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:20 crc kubenswrapper[5025]: E1004 10:35:20.411216 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:20 crc kubenswrapper[5025]: E1004 10:35:20.411475 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.482540 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.482581 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.482590 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.482605 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.482615 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:20Z","lastTransitionTime":"2025-10-04T10:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.585046 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.585095 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.585105 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.585122 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.585133 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:20Z","lastTransitionTime":"2025-10-04T10:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.688363 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.688417 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.688432 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.688456 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.688475 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:20Z","lastTransitionTime":"2025-10-04T10:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.790716 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.790754 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.790768 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.790783 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.790792 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:20Z","lastTransitionTime":"2025-10-04T10:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.845998 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/2.log" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.893561 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.893645 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.893662 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.893684 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.893699 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:20Z","lastTransitionTime":"2025-10-04T10:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.997144 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.997221 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.997239 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.997266 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:20 crc kubenswrapper[5025]: I1004 10:35:20.997287 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:20Z","lastTransitionTime":"2025-10-04T10:35:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.099839 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.099904 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.099923 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.099948 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.099968 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:21Z","lastTransitionTime":"2025-10-04T10:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.203751 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.203822 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.203839 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.203867 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.203885 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:21Z","lastTransitionTime":"2025-10-04T10:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.307397 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.307459 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.307472 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.307494 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.307511 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:21Z","lastTransitionTime":"2025-10-04T10:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.410297 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.410396 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.410448 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.410468 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.410494 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:21 crc kubenswrapper[5025]: E1004 10:35:21.410514 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.410513 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:21Z","lastTransitionTime":"2025-10-04T10:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.513601 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.513641 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.513649 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.513662 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.513672 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:21Z","lastTransitionTime":"2025-10-04T10:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.617076 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.617165 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.617185 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.617215 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.617240 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:21Z","lastTransitionTime":"2025-10-04T10:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.720597 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.720656 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.720670 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.720689 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.720704 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:21Z","lastTransitionTime":"2025-10-04T10:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.824777 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.824866 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.824893 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.824929 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.824958 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:21Z","lastTransitionTime":"2025-10-04T10:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.929108 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.929175 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.929194 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.929222 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:21 crc kubenswrapper[5025]: I1004 10:35:21.929243 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:21Z","lastTransitionTime":"2025-10-04T10:35:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.032664 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.032727 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.032741 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.032773 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.032791 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:22Z","lastTransitionTime":"2025-10-04T10:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.135701 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.135781 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.135802 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.135835 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.135859 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:22Z","lastTransitionTime":"2025-10-04T10:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.242487 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.242567 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.242587 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.242615 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.242634 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:22Z","lastTransitionTime":"2025-10-04T10:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.346547 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.346634 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.346662 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.346693 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.346717 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:22Z","lastTransitionTime":"2025-10-04T10:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.410512 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.410648 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.410716 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:22 crc kubenswrapper[5025]: E1004 10:35:22.410889 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:22 crc kubenswrapper[5025]: E1004 10:35:22.411120 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:22 crc kubenswrapper[5025]: E1004 10:35:22.411277 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.433373 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.450557 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.450689 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.450714 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.450740 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.450759 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:22Z","lastTransitionTime":"2025-10-04T10:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.453263 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.469303 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.487711 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.510529 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.534474 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.553328 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.553384 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.553400 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.553423 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.553439 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:22Z","lastTransitionTime":"2025-10-04T10:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.557304 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.574539 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.601963 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c33bc4d5f99df8bed933d149acc397fc38788e9182d14cad6492caaa662840df\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"message\\\":\\\"il,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1004 10:35:06.985348 6462 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:06.985359 6462 lb_config.go:1031] Cluster endpoints for openshift-machine-api/machine-api-operator for network=default are: map[]\\\\nI1004 10:35:06.985373 6462 services_controller.go:443] Built service openshift-machine-api/machine-api-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.5.21\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:8443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1004 10:35:06.985386 6462 services_controller.go:444] Built service openshift-machine-api/machine-api-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1004 10:35:06.985394 6462 services_controller.go:445] Built service openshift-machine-api/machine-api-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nF1004 10:35:06.985431 6462 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.620583 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.655292 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.657106 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.657155 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.657172 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.657191 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.657203 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:22Z","lastTransitionTime":"2025-10-04T10:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.676701 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.692692 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.711350 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.729304 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.751604 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.760781 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.760829 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.760841 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.760862 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.760875 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:22Z","lastTransitionTime":"2025-10-04T10:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.767592 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:22Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.864342 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.864411 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.864431 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.864457 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.864476 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:22Z","lastTransitionTime":"2025-10-04T10:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.967306 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.967376 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.967397 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.967422 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:22 crc kubenswrapper[5025]: I1004 10:35:22.967440 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:22Z","lastTransitionTime":"2025-10-04T10:35:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.070491 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.070614 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.070647 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.070733 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.070814 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:23Z","lastTransitionTime":"2025-10-04T10:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.173568 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.174150 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.174200 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.174237 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.174264 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:23Z","lastTransitionTime":"2025-10-04T10:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.277740 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.277815 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.277852 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.277883 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.277905 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:23Z","lastTransitionTime":"2025-10-04T10:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.380715 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.380817 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.380837 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.380862 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.380881 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:23Z","lastTransitionTime":"2025-10-04T10:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.411277 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:23 crc kubenswrapper[5025]: E1004 10:35:23.411537 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.420317 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:23 crc kubenswrapper[5025]: E1004 10:35:23.420657 5025 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:23 crc kubenswrapper[5025]: E1004 10:35:23.420751 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs podName:86312d38-63d0-409a-98d4-727f0fb47929 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:39.42072612 +0000 UTC m=+67.845693000 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs") pod "network-metrics-daemon-frc27" (UID: "86312d38-63d0-409a-98d4-727f0fb47929") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.483926 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.484009 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.484068 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.484093 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.484111 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:23Z","lastTransitionTime":"2025-10-04T10:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.587582 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.587657 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.587680 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.587708 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.587729 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:23Z","lastTransitionTime":"2025-10-04T10:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.690145 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.690198 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.690208 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.690228 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.690238 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:23Z","lastTransitionTime":"2025-10-04T10:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.793263 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.793316 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.793326 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.793343 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.793357 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:23Z","lastTransitionTime":"2025-10-04T10:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.896091 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.896136 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.896147 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.896164 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.896175 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:23Z","lastTransitionTime":"2025-10-04T10:35:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.953071 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.955167 5025 scope.go:117] "RemoveContainer" containerID="161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc" Oct 04 10:35:23 crc kubenswrapper[5025]: E1004 10:35:23.955444 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.977470 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:23 crc kubenswrapper[5025]: I1004 10:35:23.995398 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:23Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.001202 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.001260 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.001275 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.001341 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.001358 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:24Z","lastTransitionTime":"2025-10-04T10:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.013446 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.037582 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.070378 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.102893 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.104228 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.104294 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.104311 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.104335 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.104351 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:24Z","lastTransitionTime":"2025-10-04T10:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.121750 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.128749 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.128923 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:35:56.128892915 +0000 UTC m=+84.553859795 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.128989 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.129056 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.129095 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.129144 5025 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.129204 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:56.129189874 +0000 UTC m=+84.554156754 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.129233 5025 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.129277 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:56.129269847 +0000 UTC m=+84.554236727 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.129338 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.129389 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.129406 5025 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.129482 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:56.129459772 +0000 UTC m=+84.554426652 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.135554 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.154799 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.169381 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.183251 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.196236 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.206781 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.206826 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.206837 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.206855 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.206867 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:24Z","lastTransitionTime":"2025-10-04T10:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.211417 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.226165 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.229903 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.230112 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.230157 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.230181 5025 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.230259 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:35:56.230236686 +0000 UTC m=+84.655203606 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.236768 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.260651 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.272124 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:24Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.310706 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.310772 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.310785 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.310811 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.310825 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:24Z","lastTransitionTime":"2025-10-04T10:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.410346 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.410357 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.410502 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.410535 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.410739 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:24 crc kubenswrapper[5025]: E1004 10:35:24.410715 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.413927 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.414008 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.414068 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.414100 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.414126 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:24Z","lastTransitionTime":"2025-10-04T10:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.517250 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.517296 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.517305 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.517324 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.517336 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:24Z","lastTransitionTime":"2025-10-04T10:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.620572 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.620626 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.620637 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.620654 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.620666 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:24Z","lastTransitionTime":"2025-10-04T10:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.724253 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.724685 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.724913 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.725189 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.725368 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:24Z","lastTransitionTime":"2025-10-04T10:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.828380 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.828800 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.828957 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.829152 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.829315 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:24Z","lastTransitionTime":"2025-10-04T10:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.931581 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.931620 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.931631 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.931650 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:24 crc kubenswrapper[5025]: I1004 10:35:24.931659 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:24Z","lastTransitionTime":"2025-10-04T10:35:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.034629 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.034692 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.034711 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.034737 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.034755 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:25Z","lastTransitionTime":"2025-10-04T10:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.137828 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.137912 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.137929 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.137952 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.137971 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:25Z","lastTransitionTime":"2025-10-04T10:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.240516 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.240582 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.240605 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.240639 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.240669 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:25Z","lastTransitionTime":"2025-10-04T10:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.343849 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.343912 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.343931 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.343957 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.343979 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:25Z","lastTransitionTime":"2025-10-04T10:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.410859 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:25 crc kubenswrapper[5025]: E1004 10:35:25.411112 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.446566 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.446615 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.446628 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.446646 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.446659 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:25Z","lastTransitionTime":"2025-10-04T10:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.549780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.549836 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.549857 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.549879 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.549897 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:25Z","lastTransitionTime":"2025-10-04T10:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.653082 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.653143 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.653163 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.653187 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.653206 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:25Z","lastTransitionTime":"2025-10-04T10:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.755972 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.756074 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.756103 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.756143 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.756168 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:25Z","lastTransitionTime":"2025-10-04T10:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.859278 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.859346 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.859364 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.859389 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.859410 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:25Z","lastTransitionTime":"2025-10-04T10:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.962780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.962869 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.962896 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.962930 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:25 crc kubenswrapper[5025]: I1004 10:35:25.962955 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:25Z","lastTransitionTime":"2025-10-04T10:35:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.066073 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.066166 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.066180 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.066198 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.066210 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:26Z","lastTransitionTime":"2025-10-04T10:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.169976 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.170056 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.170067 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.170088 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.170103 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:26Z","lastTransitionTime":"2025-10-04T10:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.273121 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.273195 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.273210 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.273234 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.273262 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:26Z","lastTransitionTime":"2025-10-04T10:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.377004 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.377120 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.377145 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.377177 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.377200 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:26Z","lastTransitionTime":"2025-10-04T10:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.410953 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.410953 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:26 crc kubenswrapper[5025]: E1004 10:35:26.411155 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.410974 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:26 crc kubenswrapper[5025]: E1004 10:35:26.411289 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:26 crc kubenswrapper[5025]: E1004 10:35:26.411489 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.480411 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.480464 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.480479 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.480501 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.480520 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:26Z","lastTransitionTime":"2025-10-04T10:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.583348 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.583412 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.583428 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.583450 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.583464 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:26Z","lastTransitionTime":"2025-10-04T10:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.686291 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.686325 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.686333 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.686349 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.686359 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:26Z","lastTransitionTime":"2025-10-04T10:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.789837 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.789892 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.789917 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.789946 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.789968 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:26Z","lastTransitionTime":"2025-10-04T10:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.892847 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.892929 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.892950 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.892975 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.892997 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:26Z","lastTransitionTime":"2025-10-04T10:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.995615 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.995679 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.995697 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.995721 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:26 crc kubenswrapper[5025]: I1004 10:35:26.995739 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:26Z","lastTransitionTime":"2025-10-04T10:35:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.099506 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.099575 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.099592 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.099619 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.099638 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:27Z","lastTransitionTime":"2025-10-04T10:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.201587 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.201628 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.201639 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.201657 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.201670 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:27Z","lastTransitionTime":"2025-10-04T10:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.304373 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.304415 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.304426 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.304443 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.304453 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:27Z","lastTransitionTime":"2025-10-04T10:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.406365 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.406406 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.406415 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.406431 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.406442 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:27Z","lastTransitionTime":"2025-10-04T10:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.410772 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:27 crc kubenswrapper[5025]: E1004 10:35:27.410879 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.509761 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.509825 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.509842 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.509867 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.509881 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:27Z","lastTransitionTime":"2025-10-04T10:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.613087 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.613163 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.613184 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.613210 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.613229 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:27Z","lastTransitionTime":"2025-10-04T10:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.716692 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.716756 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.716777 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.716808 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.716828 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:27Z","lastTransitionTime":"2025-10-04T10:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.819992 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.820119 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.820144 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.820177 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.820207 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:27Z","lastTransitionTime":"2025-10-04T10:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.923223 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.923310 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.923337 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.923370 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:27 crc kubenswrapper[5025]: I1004 10:35:27.923397 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:27Z","lastTransitionTime":"2025-10-04T10:35:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.026784 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.026854 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.026872 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.026902 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.026922 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.130342 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.130422 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.130446 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.130478 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.130505 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.233293 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.233364 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.233376 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.233414 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.233430 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.330636 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.335693 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.335726 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.335734 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.335748 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.335757 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.342555 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.345308 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.363191 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.382610 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.403089 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.410210 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.410265 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:28 crc kubenswrapper[5025]: E1004 10:35:28.410361 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.410287 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:28 crc kubenswrapper[5025]: E1004 10:35:28.410516 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:28 crc kubenswrapper[5025]: E1004 10:35:28.410670 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.422813 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.440849 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.440912 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.440931 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.440957 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.440977 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.465179 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.483049 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.502859 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.522582 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.540838 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.544663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.544747 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.544772 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.544805 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.544831 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.576277 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.593383 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.593453 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.593471 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.593498 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.593516 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.600291 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: E1004 10:35:28.615192 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.618892 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.621276 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.621333 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.621350 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.621377 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.621397 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.642571 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: E1004 10:35:28.645438 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.651231 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.651428 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.651584 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.651728 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.651850 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.669106 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: E1004 10:35:28.671880 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.678047 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.678112 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.678131 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.678158 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.678178 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.690127 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: E1004 10:35:28.696450 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.701505 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.701562 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.701583 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.701605 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.701622 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.714795 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: E1004 10:35:28.721452 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:28Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:28 crc kubenswrapper[5025]: E1004 10:35:28.721687 5025 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.723906 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.723983 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.724003 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.724058 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.724082 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.827656 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.827727 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.827781 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.827810 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.827832 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.931609 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.931683 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.931708 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.931740 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:28 crc kubenswrapper[5025]: I1004 10:35:28.931757 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:28Z","lastTransitionTime":"2025-10-04T10:35:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.034879 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.034938 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.034955 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.034980 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.035001 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:29Z","lastTransitionTime":"2025-10-04T10:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.138693 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.138844 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.138874 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.138907 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.138929 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:29Z","lastTransitionTime":"2025-10-04T10:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.242539 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.242624 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.242651 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.242682 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.242707 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:29Z","lastTransitionTime":"2025-10-04T10:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.347064 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.347138 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.347158 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.347184 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.347203 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:29Z","lastTransitionTime":"2025-10-04T10:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.410858 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:29 crc kubenswrapper[5025]: E1004 10:35:29.411122 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.450260 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.450311 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.450325 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.450346 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.450361 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:29Z","lastTransitionTime":"2025-10-04T10:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.556854 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.556921 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.556934 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.556967 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.556978 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:29Z","lastTransitionTime":"2025-10-04T10:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.660161 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.660207 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.660220 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.660237 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.660248 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:29Z","lastTransitionTime":"2025-10-04T10:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.764996 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.765120 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.765141 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.765167 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.765185 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:29Z","lastTransitionTime":"2025-10-04T10:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.869355 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.869502 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.869533 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.869566 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.869589 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:29Z","lastTransitionTime":"2025-10-04T10:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.973771 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.973859 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.973879 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.973907 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:29 crc kubenswrapper[5025]: I1004 10:35:29.973926 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:29Z","lastTransitionTime":"2025-10-04T10:35:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.077450 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.077517 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.077534 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.077560 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.077579 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:30Z","lastTransitionTime":"2025-10-04T10:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.181649 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.181726 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.181745 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.181771 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.181791 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:30Z","lastTransitionTime":"2025-10-04T10:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.285169 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.285250 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.285268 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.285295 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.285314 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:30Z","lastTransitionTime":"2025-10-04T10:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.388645 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.388731 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.388750 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.388776 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.388793 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:30Z","lastTransitionTime":"2025-10-04T10:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.410756 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.410843 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.410855 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:30 crc kubenswrapper[5025]: E1004 10:35:30.410995 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:30 crc kubenswrapper[5025]: E1004 10:35:30.411157 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:30 crc kubenswrapper[5025]: E1004 10:35:30.411334 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.491409 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.491476 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.491494 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.491529 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.491547 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:30Z","lastTransitionTime":"2025-10-04T10:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.594242 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.594372 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.594455 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.594488 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.594507 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:30Z","lastTransitionTime":"2025-10-04T10:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.697854 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.697930 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.697954 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.697985 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.698007 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:30Z","lastTransitionTime":"2025-10-04T10:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.801872 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.801945 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.801968 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.802000 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.802068 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:30Z","lastTransitionTime":"2025-10-04T10:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.905068 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.905156 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.905183 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.905215 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:30 crc kubenswrapper[5025]: I1004 10:35:30.905240 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:30Z","lastTransitionTime":"2025-10-04T10:35:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.008324 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.008388 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.008408 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.008436 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.008456 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:31Z","lastTransitionTime":"2025-10-04T10:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.111558 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.111622 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.111641 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.111668 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.111687 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:31Z","lastTransitionTime":"2025-10-04T10:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.214688 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.214767 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.214788 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.214814 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.214835 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:31Z","lastTransitionTime":"2025-10-04T10:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.318199 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.318280 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.318310 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.318342 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.318366 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:31Z","lastTransitionTime":"2025-10-04T10:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.410928 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:31 crc kubenswrapper[5025]: E1004 10:35:31.411199 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.422857 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.422937 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.422967 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.422998 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.423053 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:31Z","lastTransitionTime":"2025-10-04T10:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.526597 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.526673 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.526689 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.526713 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.526731 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:31Z","lastTransitionTime":"2025-10-04T10:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.630301 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.630423 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.630442 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.630474 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.630495 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:31Z","lastTransitionTime":"2025-10-04T10:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.733961 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.734066 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.734091 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.734121 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.734140 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:31Z","lastTransitionTime":"2025-10-04T10:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.838052 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.838129 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.838148 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.838174 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.838193 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:31Z","lastTransitionTime":"2025-10-04T10:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.941517 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.941569 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.941581 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.941599 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:31 crc kubenswrapper[5025]: I1004 10:35:31.941610 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:31Z","lastTransitionTime":"2025-10-04T10:35:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.045284 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.045333 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.045344 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.045363 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.045378 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:32Z","lastTransitionTime":"2025-10-04T10:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.149444 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.149522 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.149541 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.149570 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.149592 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:32Z","lastTransitionTime":"2025-10-04T10:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.253498 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.253574 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.253594 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.253621 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.253644 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:32Z","lastTransitionTime":"2025-10-04T10:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.357504 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.357572 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.357589 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.357614 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.357632 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:32Z","lastTransitionTime":"2025-10-04T10:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.410273 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.410397 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:32 crc kubenswrapper[5025]: E1004 10:35:32.410468 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.410402 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:32 crc kubenswrapper[5025]: E1004 10:35:32.410685 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:32 crc kubenswrapper[5025]: E1004 10:35:32.410667 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.430284 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.452261 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a4bf1ab-6621-4e87-a088-139aaa7c8406\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf053ec8fd21d01980e6f7f6d5f24328ff39cf7d641f4ed959ce0ca034f4fbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f018adfdcab8bd60cd03879d0807145bf0e8b398a9a138739387f0f420c7989b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6bc1b4dcad8484ee9c6f82153ed8d93448d8c3540d44e6990981983f9573cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.460355 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.460407 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.460424 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.460452 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.460472 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:32Z","lastTransitionTime":"2025-10-04T10:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.473113 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.489969 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.508497 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.531621 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.550447 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.563795 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.563862 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.563878 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.563903 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.563922 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:32Z","lastTransitionTime":"2025-10-04T10:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.569739 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.586044 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.598709 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.617471 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.632632 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.655416 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.667169 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.667225 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.667237 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.667258 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.667273 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:32Z","lastTransitionTime":"2025-10-04T10:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.670705 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.695475 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.728082 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.749699 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.768677 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:32Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.771537 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.771614 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.771634 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.771663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.771712 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:32Z","lastTransitionTime":"2025-10-04T10:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.874596 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.874648 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.874660 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.874677 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.874688 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:32Z","lastTransitionTime":"2025-10-04T10:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.977123 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.977193 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.977211 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.977238 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:32 crc kubenswrapper[5025]: I1004 10:35:32.977260 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:32Z","lastTransitionTime":"2025-10-04T10:35:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.080418 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.080552 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.080574 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.080600 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.080659 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:33Z","lastTransitionTime":"2025-10-04T10:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.183407 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.183462 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.183479 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.183503 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.183523 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:33Z","lastTransitionTime":"2025-10-04T10:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.286827 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.286882 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.286894 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.286914 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.286927 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:33Z","lastTransitionTime":"2025-10-04T10:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.390755 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.390813 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.390831 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.390855 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.390872 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:33Z","lastTransitionTime":"2025-10-04T10:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.411295 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:33 crc kubenswrapper[5025]: E1004 10:35:33.411694 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.493108 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.493187 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.493203 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.493224 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.493238 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:33Z","lastTransitionTime":"2025-10-04T10:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.596584 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.596687 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.596704 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.596727 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.596742 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:33Z","lastTransitionTime":"2025-10-04T10:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.699738 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.699796 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.699813 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.699839 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.699857 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:33Z","lastTransitionTime":"2025-10-04T10:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.802687 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.802768 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.802793 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.802820 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.802839 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:33Z","lastTransitionTime":"2025-10-04T10:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.905800 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.905862 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.905881 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.905906 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:33 crc kubenswrapper[5025]: I1004 10:35:33.905924 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:33Z","lastTransitionTime":"2025-10-04T10:35:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.010758 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.010840 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.010876 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.010908 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.010930 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:34Z","lastTransitionTime":"2025-10-04T10:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.114993 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.115122 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.115144 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.115173 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.115191 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:34Z","lastTransitionTime":"2025-10-04T10:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.218581 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.218641 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.218650 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.218683 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.218698 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:34Z","lastTransitionTime":"2025-10-04T10:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.321709 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.321761 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.321788 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.321810 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.321821 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:34Z","lastTransitionTime":"2025-10-04T10:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.411145 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.411226 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:34 crc kubenswrapper[5025]: E1004 10:35:34.411308 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:34 crc kubenswrapper[5025]: E1004 10:35:34.411422 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.411530 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:34 crc kubenswrapper[5025]: E1004 10:35:34.411592 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.424781 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.424820 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.424829 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.424872 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.424887 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:34Z","lastTransitionTime":"2025-10-04T10:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.527639 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.527694 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.527706 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.527727 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.527741 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:34Z","lastTransitionTime":"2025-10-04T10:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.631724 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.631791 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.631809 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.631835 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.631853 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:34Z","lastTransitionTime":"2025-10-04T10:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.736068 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.736161 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.736189 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.736223 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.736247 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:34Z","lastTransitionTime":"2025-10-04T10:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.839802 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.839854 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.839864 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.839881 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.839893 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:34Z","lastTransitionTime":"2025-10-04T10:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.943452 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.943518 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.943531 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.943551 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:34 crc kubenswrapper[5025]: I1004 10:35:34.943563 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:34Z","lastTransitionTime":"2025-10-04T10:35:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.047368 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.047451 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.047468 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.047495 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.047519 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:35Z","lastTransitionTime":"2025-10-04T10:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.151219 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.151311 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.151336 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.151366 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.151563 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:35Z","lastTransitionTime":"2025-10-04T10:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.255904 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.255985 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.256005 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.256064 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.256088 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:35Z","lastTransitionTime":"2025-10-04T10:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.360531 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.360646 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.360667 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.360761 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.360828 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:35Z","lastTransitionTime":"2025-10-04T10:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.411426 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:35 crc kubenswrapper[5025]: E1004 10:35:35.411699 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.465291 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.465372 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.465395 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.465419 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.465439 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:35Z","lastTransitionTime":"2025-10-04T10:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.567607 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.567650 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.567660 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.567676 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.567686 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:35Z","lastTransitionTime":"2025-10-04T10:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.670563 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.670635 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.670653 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.670704 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.670722 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:35Z","lastTransitionTime":"2025-10-04T10:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.773845 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.773908 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.773920 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.773942 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.773952 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:35Z","lastTransitionTime":"2025-10-04T10:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.877156 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.877229 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.877242 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.877257 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.877267 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:35Z","lastTransitionTime":"2025-10-04T10:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.980537 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.980593 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.980607 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.980629 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:35 crc kubenswrapper[5025]: I1004 10:35:35.980648 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:35Z","lastTransitionTime":"2025-10-04T10:35:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.083414 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.083479 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.083495 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.083521 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.083539 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:36Z","lastTransitionTime":"2025-10-04T10:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.187013 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.187061 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.187071 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.187085 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.187096 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:36Z","lastTransitionTime":"2025-10-04T10:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.290177 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.290282 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.290306 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.290336 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.290364 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:36Z","lastTransitionTime":"2025-10-04T10:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.393348 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.393404 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.393416 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.393436 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.393449 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:36Z","lastTransitionTime":"2025-10-04T10:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.410822 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.410892 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:36 crc kubenswrapper[5025]: E1004 10:35:36.410988 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.411183 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:36 crc kubenswrapper[5025]: E1004 10:35:36.411346 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:36 crc kubenswrapper[5025]: E1004 10:35:36.411484 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.496879 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.496957 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.496976 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.497004 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.497048 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:36Z","lastTransitionTime":"2025-10-04T10:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.600394 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.600480 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.600496 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.600519 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.600555 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:36Z","lastTransitionTime":"2025-10-04T10:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.703849 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.703901 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.703916 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.703934 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.703950 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:36Z","lastTransitionTime":"2025-10-04T10:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.806875 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.806920 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.806929 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.806947 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.806958 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:36Z","lastTransitionTime":"2025-10-04T10:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.910118 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.910177 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.910192 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.910213 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:36 crc kubenswrapper[5025]: I1004 10:35:36.910227 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:36Z","lastTransitionTime":"2025-10-04T10:35:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.013206 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.013266 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.013287 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.013315 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.013334 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:37Z","lastTransitionTime":"2025-10-04T10:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.115870 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.115924 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.115933 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.115952 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.115969 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:37Z","lastTransitionTime":"2025-10-04T10:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.218033 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.218070 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.218079 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.218093 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.218103 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:37Z","lastTransitionTime":"2025-10-04T10:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.321180 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.321278 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.321302 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.321335 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.321365 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:37Z","lastTransitionTime":"2025-10-04T10:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.410577 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:37 crc kubenswrapper[5025]: E1004 10:35:37.410785 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.423921 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.424005 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.424039 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.424064 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.424077 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:37Z","lastTransitionTime":"2025-10-04T10:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.527668 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.527751 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.527778 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.527810 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.527832 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:37Z","lastTransitionTime":"2025-10-04T10:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.631971 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.632059 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.632075 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.632097 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.632114 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:37Z","lastTransitionTime":"2025-10-04T10:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.734688 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.734766 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.734792 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.734820 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.734839 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:37Z","lastTransitionTime":"2025-10-04T10:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.838824 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.838890 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.838908 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.838936 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.838954 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:37Z","lastTransitionTime":"2025-10-04T10:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.941195 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.941252 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.941271 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.941296 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:37 crc kubenswrapper[5025]: I1004 10:35:37.941317 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:37Z","lastTransitionTime":"2025-10-04T10:35:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.044487 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.044562 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.044588 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.044624 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.044654 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.147417 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.147489 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.147510 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.147537 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.147556 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.249935 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.249984 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.249996 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.250031 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.250064 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.353110 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.353158 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.353170 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.353187 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.353200 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.410764 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:38 crc kubenswrapper[5025]: E1004 10:35:38.410891 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.410785 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:38 crc kubenswrapper[5025]: E1004 10:35:38.410976 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.410764 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:38 crc kubenswrapper[5025]: E1004 10:35:38.411604 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.412376 5025 scope.go:117] "RemoveContainer" containerID="161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc" Oct 04 10:35:38 crc kubenswrapper[5025]: E1004 10:35:38.412763 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.455467 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.455513 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.455524 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.455542 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.455557 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.558435 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.558492 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.558509 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.558536 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.558553 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.661849 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.661913 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.661926 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.661943 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.661958 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.765722 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.765762 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.765797 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.765819 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.765831 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.868732 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.868774 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.868785 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.868799 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.868809 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.869869 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.869936 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.869954 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.869983 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.870001 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: E1004 10:35:38.884929 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:38Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.888974 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.889053 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.889070 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.889088 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.889124 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: E1004 10:35:38.899560 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:38Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.903737 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.903770 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.903781 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.903801 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.903812 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: E1004 10:35:38.915850 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:38Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.919285 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.919348 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.919369 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.919395 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.919417 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: E1004 10:35:38.931872 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:38Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.935037 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.935072 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.935083 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.935102 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.935114 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:38 crc kubenswrapper[5025]: E1004 10:35:38.945413 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:38Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:38 crc kubenswrapper[5025]: E1004 10:35:38.945548 5025 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.971185 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.971222 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.971233 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.971253 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:38 crc kubenswrapper[5025]: I1004 10:35:38.971265 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:38Z","lastTransitionTime":"2025-10-04T10:35:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.074222 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.074263 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.074280 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.074304 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.074323 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:39Z","lastTransitionTime":"2025-10-04T10:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.177581 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.177624 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.177638 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.177656 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.177670 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:39Z","lastTransitionTime":"2025-10-04T10:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.280521 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.280569 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.280579 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.280597 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.280607 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:39Z","lastTransitionTime":"2025-10-04T10:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.382839 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.382893 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.382908 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.382928 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.382941 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:39Z","lastTransitionTime":"2025-10-04T10:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.410572 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:39 crc kubenswrapper[5025]: E1004 10:35:39.410717 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.485707 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.485752 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.485764 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.485780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.485792 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:39Z","lastTransitionTime":"2025-10-04T10:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.519078 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:39 crc kubenswrapper[5025]: E1004 10:35:39.519261 5025 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:39 crc kubenswrapper[5025]: E1004 10:35:39.519348 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs podName:86312d38-63d0-409a-98d4-727f0fb47929 nodeName:}" failed. No retries permitted until 2025-10-04 10:36:11.519325544 +0000 UTC m=+99.944292454 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs") pod "network-metrics-daemon-frc27" (UID: "86312d38-63d0-409a-98d4-727f0fb47929") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.588792 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.588863 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.588882 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.588911 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.588931 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:39Z","lastTransitionTime":"2025-10-04T10:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.692303 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.692369 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.692387 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.692414 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.692436 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:39Z","lastTransitionTime":"2025-10-04T10:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.795871 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.795926 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.795942 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.795964 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.795975 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:39Z","lastTransitionTime":"2025-10-04T10:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.898632 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.898681 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.898692 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.898742 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:39 crc kubenswrapper[5025]: I1004 10:35:39.898774 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:39Z","lastTransitionTime":"2025-10-04T10:35:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.002966 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.003056 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.003071 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.003091 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.003130 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:40Z","lastTransitionTime":"2025-10-04T10:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.106624 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.106693 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.106712 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.106739 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.106759 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:40Z","lastTransitionTime":"2025-10-04T10:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.209620 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.209690 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.209708 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.209728 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.209741 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:40Z","lastTransitionTime":"2025-10-04T10:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.312484 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.312564 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.312583 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.312612 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.312631 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:40Z","lastTransitionTime":"2025-10-04T10:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.411255 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.411353 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:40 crc kubenswrapper[5025]: E1004 10:35:40.411449 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.411491 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:40 crc kubenswrapper[5025]: E1004 10:35:40.411541 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:40 crc kubenswrapper[5025]: E1004 10:35:40.411668 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.420339 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.420453 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.420491 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.420516 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.420535 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:40Z","lastTransitionTime":"2025-10-04T10:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.525299 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.525352 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.525369 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.525392 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.525410 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:40Z","lastTransitionTime":"2025-10-04T10:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.627982 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.628037 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.628048 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.628066 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.628079 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:40Z","lastTransitionTime":"2025-10-04T10:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.730547 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.730579 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.730587 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.730601 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.730610 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:40Z","lastTransitionTime":"2025-10-04T10:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.833583 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.833631 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.833641 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.833657 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.833667 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:40Z","lastTransitionTime":"2025-10-04T10:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.927401 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n49fv_8d214492-7532-4f72-b032-b3cc1ae19473/kube-multus/0.log" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.927454 5025 generic.go:334] "Generic (PLEG): container finished" podID="8d214492-7532-4f72-b032-b3cc1ae19473" containerID="507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172" exitCode=1 Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.927488 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n49fv" event={"ID":"8d214492-7532-4f72-b032-b3cc1ae19473","Type":"ContainerDied","Data":"507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.927954 5025 scope.go:117] "RemoveContainer" containerID="507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.936368 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.936413 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.936425 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.936442 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.936457 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:40Z","lastTransitionTime":"2025-10-04T10:35:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.944161 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:40Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.959685 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:40Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.974892 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:40Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:40 crc kubenswrapper[5025]: I1004 10:35:40.990956 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:40Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:40Z\\\",\\\"message\\\":\\\"2025-10-04T10:34:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413\\\\n2025-10-04T10:34:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413 to /host/opt/cni/bin/\\\\n2025-10-04T10:34:55Z [verbose] multus-daemon started\\\\n2025-10-04T10:34:55Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:35:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:40Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.004038 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.028749 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.039207 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.039238 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.039246 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.039260 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.039270 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:41Z","lastTransitionTime":"2025-10-04T10:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.050716 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.070420 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.082125 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.093540 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.111694 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a4bf1ab-6621-4e87-a088-139aaa7c8406\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf053ec8fd21d01980e6f7f6d5f24328ff39cf7d641f4ed959ce0ca034f4fbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f018adfdcab8bd60cd03879d0807145bf0e8b398a9a138739387f0f420c7989b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6bc1b4dcad8484ee9c6f82153ed8d93448d8c3540d44e6990981983f9573cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.125991 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.141634 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.142462 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.142493 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.142505 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.142520 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.142531 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:41Z","lastTransitionTime":"2025-10-04T10:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.163514 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.176236 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.187362 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.197254 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.206314 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.244247 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.244282 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.244294 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.244310 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.244322 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:41Z","lastTransitionTime":"2025-10-04T10:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.346359 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.346383 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.346391 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.346405 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.346414 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:41Z","lastTransitionTime":"2025-10-04T10:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.411129 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:41 crc kubenswrapper[5025]: E1004 10:35:41.411277 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.447954 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.447995 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.448005 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.448054 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.448068 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:41Z","lastTransitionTime":"2025-10-04T10:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.550818 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.550859 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.550870 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.550886 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.550898 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:41Z","lastTransitionTime":"2025-10-04T10:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.654420 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.654503 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.654519 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.654541 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.654554 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:41Z","lastTransitionTime":"2025-10-04T10:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.757522 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.757585 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.757632 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.757657 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.757675 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:41Z","lastTransitionTime":"2025-10-04T10:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.860549 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.860613 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.860632 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.860655 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.860670 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:41Z","lastTransitionTime":"2025-10-04T10:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.935347 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n49fv_8d214492-7532-4f72-b032-b3cc1ae19473/kube-multus/0.log" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.936181 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n49fv" event={"ID":"8d214492-7532-4f72-b032-b3cc1ae19473","Type":"ContainerStarted","Data":"30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.964487 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.964548 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.964566 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.964597 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.964672 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:41Z","lastTransitionTime":"2025-10-04T10:35:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.964722 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.977299 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:41 crc kubenswrapper[5025]: I1004 10:35:41.989058 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:41Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.009989 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.023046 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.036595 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a4bf1ab-6621-4e87-a088-139aaa7c8406\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf053ec8fd21d01980e6f7f6d5f24328ff39cf7d641f4ed959ce0ca034f4fbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f018adfdcab8bd60cd03879d0807145bf0e8b398a9a138739387f0f420c7989b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6bc1b4dcad8484ee9c6f82153ed8d93448d8c3540d44e6990981983f9573cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.053898 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.066217 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.067110 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.067153 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.067162 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.067179 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.067191 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:42Z","lastTransitionTime":"2025-10-04T10:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.082190 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.097634 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.109904 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.120563 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.147053 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.161097 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.169599 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.169673 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.169690 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.169717 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.169740 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:42Z","lastTransitionTime":"2025-10-04T10:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.173210 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.186542 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.201178 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:40Z\\\",\\\"message\\\":\\\"2025-10-04T10:34:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413\\\\n2025-10-04T10:34:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413 to /host/opt/cni/bin/\\\\n2025-10-04T10:34:55Z [verbose] multus-daemon started\\\\n2025-10-04T10:34:55Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:35:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.212267 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.274066 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.274173 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.274195 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.274225 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.274244 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:42Z","lastTransitionTime":"2025-10-04T10:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.383913 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.383975 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.383992 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.384043 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.384063 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:42Z","lastTransitionTime":"2025-10-04T10:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.411150 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.411292 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:42 crc kubenswrapper[5025]: E1004 10:35:42.411503 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.411668 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:42 crc kubenswrapper[5025]: E1004 10:35:42.411842 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:42 crc kubenswrapper[5025]: E1004 10:35:42.411997 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.431066 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.447496 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.459070 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.471330 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.486616 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.486641 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.486649 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.486665 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.486675 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:42Z","lastTransitionTime":"2025-10-04T10:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.490224 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.506322 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:40Z\\\",\\\"message\\\":\\\"2025-10-04T10:34:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413\\\\n2025-10-04T10:34:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413 to /host/opt/cni/bin/\\\\n2025-10-04T10:34:55Z [verbose] multus-daemon started\\\\n2025-10-04T10:34:55Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:35:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.517877 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.542307 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.556242 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.570548 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.584726 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.590702 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.590804 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.603151 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.603246 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.603324 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:42Z","lastTransitionTime":"2025-10-04T10:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.611577 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.654778 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.677149 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.693077 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.706615 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.706678 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.706688 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.706709 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.706720 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:42Z","lastTransitionTime":"2025-10-04T10:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.708661 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.721004 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.733796 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a4bf1ab-6621-4e87-a088-139aaa7c8406\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf053ec8fd21d01980e6f7f6d5f24328ff39cf7d641f4ed959ce0ca034f4fbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f018adfdcab8bd60cd03879d0807145bf0e8b398a9a138739387f0f420c7989b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6bc1b4dcad8484ee9c6f82153ed8d93448d8c3540d44e6990981983f9573cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:42Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.809190 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.809260 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.809277 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.809303 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.809323 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:42Z","lastTransitionTime":"2025-10-04T10:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.912447 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.912522 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.912540 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.912602 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:42 crc kubenswrapper[5025]: I1004 10:35:42.912621 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:42Z","lastTransitionTime":"2025-10-04T10:35:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.015068 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.015137 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.015151 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.015172 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.015186 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:43Z","lastTransitionTime":"2025-10-04T10:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.116947 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.117004 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.117073 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.117094 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.117106 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:43Z","lastTransitionTime":"2025-10-04T10:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.219986 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.220065 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.220075 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.220092 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.220101 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:43Z","lastTransitionTime":"2025-10-04T10:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.322648 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.322705 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.322717 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.322736 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.322748 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:43Z","lastTransitionTime":"2025-10-04T10:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.410809 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:43 crc kubenswrapper[5025]: E1004 10:35:43.410931 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.425668 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.425733 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.425757 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.425816 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.425837 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:43Z","lastTransitionTime":"2025-10-04T10:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.527991 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.528040 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.528050 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.528064 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.528075 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:43Z","lastTransitionTime":"2025-10-04T10:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.630765 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.630810 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.630821 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.630841 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.630854 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:43Z","lastTransitionTime":"2025-10-04T10:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.734108 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.734176 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.734239 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.734269 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.734288 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:43Z","lastTransitionTime":"2025-10-04T10:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.837471 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.837536 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.837554 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.837581 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.837600 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:43Z","lastTransitionTime":"2025-10-04T10:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.941348 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.941426 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.941440 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.941462 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:43 crc kubenswrapper[5025]: I1004 10:35:43.941474 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:43Z","lastTransitionTime":"2025-10-04T10:35:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.044840 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.044889 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.044898 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.044919 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.044930 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:44Z","lastTransitionTime":"2025-10-04T10:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.147815 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.147896 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.147922 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.147957 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.147982 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:44Z","lastTransitionTime":"2025-10-04T10:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.251561 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.251682 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.251704 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.251736 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.251758 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:44Z","lastTransitionTime":"2025-10-04T10:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.354553 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.354608 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.354627 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.354652 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.354667 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:44Z","lastTransitionTime":"2025-10-04T10:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.410729 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.410792 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:44 crc kubenswrapper[5025]: E1004 10:35:44.410882 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.410730 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:44 crc kubenswrapper[5025]: E1004 10:35:44.410992 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:44 crc kubenswrapper[5025]: E1004 10:35:44.411270 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.457630 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.457693 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.457716 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.457745 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.457769 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:44Z","lastTransitionTime":"2025-10-04T10:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.560586 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.560648 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.560663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.560684 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.560700 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:44Z","lastTransitionTime":"2025-10-04T10:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.663644 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.663685 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.663698 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.663719 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.663729 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:44Z","lastTransitionTime":"2025-10-04T10:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.766840 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.766911 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.766934 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.766967 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.766990 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:44Z","lastTransitionTime":"2025-10-04T10:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.870741 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.870823 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.870844 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.870877 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.870899 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:44Z","lastTransitionTime":"2025-10-04T10:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.974111 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.974170 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.974186 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.974206 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:44 crc kubenswrapper[5025]: I1004 10:35:44.974218 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:44Z","lastTransitionTime":"2025-10-04T10:35:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.077079 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.077139 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.077152 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.077172 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.077187 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:45Z","lastTransitionTime":"2025-10-04T10:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.179480 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.179522 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.179533 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.179548 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.179558 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:45Z","lastTransitionTime":"2025-10-04T10:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.282110 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.282166 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.282184 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.282214 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.282241 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:45Z","lastTransitionTime":"2025-10-04T10:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.384767 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.384810 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.384821 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.384839 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.384851 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:45Z","lastTransitionTime":"2025-10-04T10:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.410684 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:45 crc kubenswrapper[5025]: E1004 10:35:45.410838 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.486973 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.487060 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.487079 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.487137 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.487153 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:45Z","lastTransitionTime":"2025-10-04T10:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.590376 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.590424 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.590435 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.590454 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.590465 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:45Z","lastTransitionTime":"2025-10-04T10:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.693482 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.693540 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.693552 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.693572 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.693585 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:45Z","lastTransitionTime":"2025-10-04T10:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.796111 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.796164 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.796179 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.796198 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.796213 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:45Z","lastTransitionTime":"2025-10-04T10:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.900428 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.900505 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.900531 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.900562 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:45 crc kubenswrapper[5025]: I1004 10:35:45.900583 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:45Z","lastTransitionTime":"2025-10-04T10:35:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.003033 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.003083 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.003094 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.003111 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.003122 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:46Z","lastTransitionTime":"2025-10-04T10:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.105908 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.105961 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.105978 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.106005 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.106048 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:46Z","lastTransitionTime":"2025-10-04T10:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.209103 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.209146 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.209156 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.209171 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.209181 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:46Z","lastTransitionTime":"2025-10-04T10:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.311970 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.311994 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.312005 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.312034 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.312042 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:46Z","lastTransitionTime":"2025-10-04T10:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.411518 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:46 crc kubenswrapper[5025]: E1004 10:35:46.411833 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.412204 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:46 crc kubenswrapper[5025]: E1004 10:35:46.412322 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.412390 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:46 crc kubenswrapper[5025]: E1004 10:35:46.412522 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.414500 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.414544 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.414569 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.414595 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.414615 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:46Z","lastTransitionTime":"2025-10-04T10:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.517458 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.517496 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.517506 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.517520 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.517531 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:46Z","lastTransitionTime":"2025-10-04T10:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.620749 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.620797 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.620808 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.620825 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.620838 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:46Z","lastTransitionTime":"2025-10-04T10:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.724067 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.724109 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.724120 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.724137 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.724146 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:46Z","lastTransitionTime":"2025-10-04T10:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.827559 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.827625 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.827640 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.827663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.827679 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:46Z","lastTransitionTime":"2025-10-04T10:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.931388 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.931436 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.931449 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.931468 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:46 crc kubenswrapper[5025]: I1004 10:35:46.931482 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:46Z","lastTransitionTime":"2025-10-04T10:35:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.034072 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.034111 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.034123 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.034140 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.034149 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:47Z","lastTransitionTime":"2025-10-04T10:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.138079 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.138148 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.138167 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.138198 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.138226 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:47Z","lastTransitionTime":"2025-10-04T10:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.241981 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.242086 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.242105 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.242134 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.242155 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:47Z","lastTransitionTime":"2025-10-04T10:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.345858 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.345925 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.345944 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.345970 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.345990 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:47Z","lastTransitionTime":"2025-10-04T10:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.410860 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:47 crc kubenswrapper[5025]: E1004 10:35:47.411091 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.450873 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.450927 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.450940 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.450963 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.450973 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:47Z","lastTransitionTime":"2025-10-04T10:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.554104 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.554144 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.554154 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.554171 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.554183 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:47Z","lastTransitionTime":"2025-10-04T10:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.658486 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.658578 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.658611 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.658647 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.658669 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:47Z","lastTransitionTime":"2025-10-04T10:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.763496 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.763609 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.763631 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.764938 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.764964 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:47Z","lastTransitionTime":"2025-10-04T10:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.869129 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.869188 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.869206 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.869231 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.869249 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:47Z","lastTransitionTime":"2025-10-04T10:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.972697 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.972756 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.972775 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.972800 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:47 crc kubenswrapper[5025]: I1004 10:35:47.972822 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:47Z","lastTransitionTime":"2025-10-04T10:35:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.076385 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.076451 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.076468 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.076493 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.076512 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:48Z","lastTransitionTime":"2025-10-04T10:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.180060 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.180122 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.180140 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.180165 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.180184 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:48Z","lastTransitionTime":"2025-10-04T10:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.283316 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.283386 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.283405 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.283432 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.283453 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:48Z","lastTransitionTime":"2025-10-04T10:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.385888 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.385924 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.385934 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.385951 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.385962 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:48Z","lastTransitionTime":"2025-10-04T10:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.410590 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.410647 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:48 crc kubenswrapper[5025]: E1004 10:35:48.410758 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.410602 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:48 crc kubenswrapper[5025]: E1004 10:35:48.410976 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:48 crc kubenswrapper[5025]: E1004 10:35:48.411078 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.489371 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.489444 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.489462 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.489490 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.489509 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:48Z","lastTransitionTime":"2025-10-04T10:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.592434 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.592556 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.592584 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.592624 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.592653 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:48Z","lastTransitionTime":"2025-10-04T10:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.695909 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.695970 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.695985 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.696007 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.696044 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:48Z","lastTransitionTime":"2025-10-04T10:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.798625 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.798707 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.798735 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.798761 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.798779 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:48Z","lastTransitionTime":"2025-10-04T10:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.902487 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.902559 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.902579 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.902605 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:48 crc kubenswrapper[5025]: I1004 10:35:48.902625 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:48Z","lastTransitionTime":"2025-10-04T10:35:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.006360 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.006403 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.006415 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.006433 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.006448 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.110478 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.110530 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.110547 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.110570 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.110588 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.213587 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.213646 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.213662 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.213688 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.213703 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.231164 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.231239 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.231262 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.231291 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.231313 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: E1004 10:35:49.254905 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.260223 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.260262 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.260278 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.260305 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.260324 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: E1004 10:35:49.280162 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.284684 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.284725 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.284744 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.284767 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.284785 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: E1004 10:35:49.304470 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.309141 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.309212 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.309230 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.309257 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.309279 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: E1004 10:35:49.329562 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.335357 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.335419 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.335445 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.335478 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.335503 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: E1004 10:35:49.358458 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:49Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:49 crc kubenswrapper[5025]: E1004 10:35:49.358833 5025 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.361600 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.361657 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.361675 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.361701 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.361719 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.410629 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:49 crc kubenswrapper[5025]: E1004 10:35:49.411405 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.465070 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.465157 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.465179 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.465212 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.465231 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.567838 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.567899 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.567919 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.567945 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.567966 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.671007 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.671080 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.671096 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.671118 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.671134 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.774450 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.774513 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.774530 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.774559 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.774582 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.877758 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.877838 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.877859 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.877893 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.877916 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.981102 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.981189 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.981229 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.981264 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:49 crc kubenswrapper[5025]: I1004 10:35:49.981289 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:49Z","lastTransitionTime":"2025-10-04T10:35:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.084451 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.084539 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.084567 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.084602 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.084626 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:50Z","lastTransitionTime":"2025-10-04T10:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.187458 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.187548 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.187571 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.187601 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.187626 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:50Z","lastTransitionTime":"2025-10-04T10:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.291109 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.291171 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.291181 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.291205 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.291229 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:50Z","lastTransitionTime":"2025-10-04T10:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.394597 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.394691 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.394718 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.394752 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.394777 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:50Z","lastTransitionTime":"2025-10-04T10:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.410547 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.410635 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:50 crc kubenswrapper[5025]: E1004 10:35:50.410815 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.411279 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:50 crc kubenswrapper[5025]: E1004 10:35:50.411601 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:50 crc kubenswrapper[5025]: E1004 10:35:50.411819 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.498939 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.498997 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.499048 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.499076 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.499094 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:50Z","lastTransitionTime":"2025-10-04T10:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.602384 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.602455 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.602471 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.602496 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.602520 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:50Z","lastTransitionTime":"2025-10-04T10:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.705539 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.705612 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.705629 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.705657 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.705675 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:50Z","lastTransitionTime":"2025-10-04T10:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.809001 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.809089 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.809107 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.809125 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.809137 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:50Z","lastTransitionTime":"2025-10-04T10:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.912805 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.912849 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.912859 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.912876 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:50 crc kubenswrapper[5025]: I1004 10:35:50.912885 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:50Z","lastTransitionTime":"2025-10-04T10:35:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.015960 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.016053 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.016077 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.016111 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.016134 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:51Z","lastTransitionTime":"2025-10-04T10:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.119824 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.119890 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.119909 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.119939 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.119957 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:51Z","lastTransitionTime":"2025-10-04T10:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.223111 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.223185 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.223212 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.223244 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.223266 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:51Z","lastTransitionTime":"2025-10-04T10:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.326496 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.326571 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.326596 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.326633 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.326659 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:51Z","lastTransitionTime":"2025-10-04T10:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.410395 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:51 crc kubenswrapper[5025]: E1004 10:35:51.410627 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.430338 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.430407 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.430428 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.430457 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.430479 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:51Z","lastTransitionTime":"2025-10-04T10:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.533735 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.533811 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.533834 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.533860 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.533879 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:51Z","lastTransitionTime":"2025-10-04T10:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.637957 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.638069 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.638090 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.638119 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.638138 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:51Z","lastTransitionTime":"2025-10-04T10:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.741251 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.741326 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.741344 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.741371 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.741391 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:51Z","lastTransitionTime":"2025-10-04T10:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.843921 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.843965 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.843975 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.843992 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.844006 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:51Z","lastTransitionTime":"2025-10-04T10:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.947098 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.947211 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.947231 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.947259 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:51 crc kubenswrapper[5025]: I1004 10:35:51.947279 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:51Z","lastTransitionTime":"2025-10-04T10:35:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.049682 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.049747 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.049767 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.049794 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.049813 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:52Z","lastTransitionTime":"2025-10-04T10:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.155159 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.155247 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.155322 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.155369 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.155392 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:52Z","lastTransitionTime":"2025-10-04T10:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.259536 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.259787 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.259805 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.259829 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.259847 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:52Z","lastTransitionTime":"2025-10-04T10:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.363583 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.363658 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.363685 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.363715 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.363735 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:52Z","lastTransitionTime":"2025-10-04T10:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.413472 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:52 crc kubenswrapper[5025]: E1004 10:35:52.413912 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.414330 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:52 crc kubenswrapper[5025]: E1004 10:35:52.414451 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.414743 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:52 crc kubenswrapper[5025]: E1004 10:35:52.416583 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.433610 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.462551 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.476629 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.476686 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.476702 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.476730 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.476749 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:52Z","lastTransitionTime":"2025-10-04T10:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.479735 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.499643 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.524273 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a4bf1ab-6621-4e87-a088-139aaa7c8406\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf053ec8fd21d01980e6f7f6d5f24328ff39cf7d641f4ed959ce0ca034f4fbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f018adfdcab8bd60cd03879d0807145bf0e8b398a9a138739387f0f420c7989b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6bc1b4dcad8484ee9c6f82153ed8d93448d8c3540d44e6990981983f9573cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.542301 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.566482 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.579635 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.579691 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.579714 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.579744 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.579770 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:52Z","lastTransitionTime":"2025-10-04T10:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.600171 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.621851 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.653401 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.675165 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.686128 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.686212 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.686591 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.686653 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.686719 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:52Z","lastTransitionTime":"2025-10-04T10:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.689554 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.703245 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.720075 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.734142 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.750596 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:40Z\\\",\\\"message\\\":\\\"2025-10-04T10:34:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413\\\\n2025-10-04T10:34:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413 to /host/opt/cni/bin/\\\\n2025-10-04T10:34:55Z [verbose] multus-daemon started\\\\n2025-10-04T10:34:55Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:35:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.764920 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.785282 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:52Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.789231 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.789271 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.789280 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.789295 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.789306 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:52Z","lastTransitionTime":"2025-10-04T10:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.891335 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.891381 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.891394 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.891414 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.891428 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:52Z","lastTransitionTime":"2025-10-04T10:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.994433 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.994524 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.994566 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.994788 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:52 crc kubenswrapper[5025]: I1004 10:35:52.994813 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:52Z","lastTransitionTime":"2025-10-04T10:35:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.097460 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.097501 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.097513 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.097530 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.097543 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:53Z","lastTransitionTime":"2025-10-04T10:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.200328 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.200390 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.200409 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.200435 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.200456 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:53Z","lastTransitionTime":"2025-10-04T10:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.303688 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.303757 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.303777 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.303804 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.303824 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:53Z","lastTransitionTime":"2025-10-04T10:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.405886 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.405926 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.405937 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.405954 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.405963 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:53Z","lastTransitionTime":"2025-10-04T10:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.411370 5025 scope.go:117] "RemoveContainer" containerID="161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.411703 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:53 crc kubenswrapper[5025]: E1004 10:35:53.411768 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.508969 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.509360 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.509373 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.509389 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.509420 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:53Z","lastTransitionTime":"2025-10-04T10:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.611901 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.611956 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.611973 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.611995 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.612061 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:53Z","lastTransitionTime":"2025-10-04T10:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.714913 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.714957 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.714968 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.714992 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.715007 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:53Z","lastTransitionTime":"2025-10-04T10:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.818006 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.818063 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.818089 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.818108 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.818118 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:53Z","lastTransitionTime":"2025-10-04T10:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.920833 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.920877 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.920888 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.920904 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.920919 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:53Z","lastTransitionTime":"2025-10-04T10:35:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.986121 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/2.log" Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.990146 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2"} Oct 04 10:35:53 crc kubenswrapper[5025]: I1004 10:35:53.990821 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.011161 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.023692 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.023726 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.023735 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.023748 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.023758 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:54Z","lastTransitionTime":"2025-10-04T10:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.031876 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.047712 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.062613 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a4bf1ab-6621-4e87-a088-139aaa7c8406\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf053ec8fd21d01980e6f7f6d5f24328ff39cf7d641f4ed959ce0ca034f4fbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f018adfdcab8bd60cd03879d0807145bf0e8b398a9a138739387f0f420c7989b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6bc1b4dcad8484ee9c6f82153ed8d93448d8c3540d44e6990981983f9573cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.077264 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.088094 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.101966 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.122130 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.126101 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.126159 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.126173 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.126197 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.126211 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:54Z","lastTransitionTime":"2025-10-04T10:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.138822 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.157493 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.195452 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.210542 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.224954 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.229145 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.229192 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.229207 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.229230 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.229246 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:54Z","lastTransitionTime":"2025-10-04T10:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.241969 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.258394 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:40Z\\\",\\\"message\\\":\\\"2025-10-04T10:34:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413\\\\n2025-10-04T10:34:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413 to /host/opt/cni/bin/\\\\n2025-10-04T10:34:55Z [verbose] multus-daemon started\\\\n2025-10-04T10:34:55Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:35:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.284648 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.315065 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.327944 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:54Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.331589 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.331618 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.331631 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.331650 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.331670 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:54Z","lastTransitionTime":"2025-10-04T10:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.410977 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.411068 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.411122 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:54 crc kubenswrapper[5025]: E1004 10:35:54.411221 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:54 crc kubenswrapper[5025]: E1004 10:35:54.411543 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:54 crc kubenswrapper[5025]: E1004 10:35:54.411401 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.434133 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.434191 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.434207 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.434230 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.434249 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:54Z","lastTransitionTime":"2025-10-04T10:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.537469 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.537527 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.537580 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.537609 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.537627 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:54Z","lastTransitionTime":"2025-10-04T10:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.640465 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.640524 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.640541 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.640566 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.640589 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:54Z","lastTransitionTime":"2025-10-04T10:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.744360 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.744434 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.744455 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.744484 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.744509 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:54Z","lastTransitionTime":"2025-10-04T10:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.847367 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.847425 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.847441 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.847465 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.847485 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:54Z","lastTransitionTime":"2025-10-04T10:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.949907 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.949962 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.949981 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.950008 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.950055 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:54Z","lastTransitionTime":"2025-10-04T10:35:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.996110 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/3.log" Oct 04 10:35:54 crc kubenswrapper[5025]: I1004 10:35:54.997205 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/2.log" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.001257 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" exitCode=1 Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.001336 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2"} Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.001422 5025 scope.go:117] "RemoveContainer" containerID="161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.002434 5025 scope.go:117] "RemoveContainer" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:35:55 crc kubenswrapper[5025]: E1004 10:35:55.002720 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.029571 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.051154 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.053938 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.054102 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.054135 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.054176 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.054203 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:55Z","lastTransitionTime":"2025-10-04T10:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.077922 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:40Z\\\",\\\"message\\\":\\\"2025-10-04T10:34:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413\\\\n2025-10-04T10:34:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413 to /host/opt/cni/bin/\\\\n2025-10-04T10:34:55Z [verbose] multus-daemon started\\\\n2025-10-04T10:34:55Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:35:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.097606 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.129751 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://161f181141ae974e62f8e417689b56adbb8007095f83e4a5ab2f347644b8fecc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:19Z\\\",\\\"message\\\":\\\"d as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1004 10:35:19.339105 6661 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1004 10:35:19.339068 6661 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1004 10:35:19.339202 6661 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:18Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:54Z\\\",\\\"message\\\":\\\"004 10:35:54.376544 7061 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 10:35:54.376630 7061 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 10:35:54.376664 7061 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 10:35:54.376847 7061 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 10:35:54.376897 7061 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 10:35:54.376905 7061 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 10:35:54.376916 7061 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 10:35:54.376922 7061 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 10:35:54.376946 7061 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 10:35:54.376963 7061 factory.go:656] Stopping watch factory\\\\nI1004 10:35:54.376974 7061 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 10:35:54.376979 7061 ovnkube.go:599] Stopped ovnkube\\\\nI1004 10:35:54.377002 7061 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 10:35:54.377008 7061 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 10:35:54.377036 7061 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 10:35:54.377030 7061 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.142119 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.156981 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.157068 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.157081 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.157099 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.157136 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:55Z","lastTransitionTime":"2025-10-04T10:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.164765 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.179775 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.193126 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.204643 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a4bf1ab-6621-4e87-a088-139aaa7c8406\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf053ec8fd21d01980e6f7f6d5f24328ff39cf7d641f4ed959ce0ca034f4fbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f018adfdcab8bd60cd03879d0807145bf0e8b398a9a138739387f0f420c7989b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6bc1b4dcad8484ee9c6f82153ed8d93448d8c3540d44e6990981983f9573cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.216278 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.228388 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.241858 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.251586 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.259556 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.259592 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.259607 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.259627 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.259641 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:55Z","lastTransitionTime":"2025-10-04T10:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.267404 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.283265 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.296659 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.307810 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:55Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.362461 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.362529 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.362554 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.362584 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.362674 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:55Z","lastTransitionTime":"2025-10-04T10:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.410772 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:55 crc kubenswrapper[5025]: E1004 10:35:55.411170 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.465996 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.466110 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.466133 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.466165 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.466184 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:55Z","lastTransitionTime":"2025-10-04T10:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.568765 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.568814 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.568827 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.568847 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.568860 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:55Z","lastTransitionTime":"2025-10-04T10:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.671933 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.672046 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.672074 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.672103 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.672130 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:55Z","lastTransitionTime":"2025-10-04T10:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.775709 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.775767 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.775785 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.775809 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.775827 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:55Z","lastTransitionTime":"2025-10-04T10:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.878777 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.879201 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.879375 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.879589 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.879728 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:55Z","lastTransitionTime":"2025-10-04T10:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.983247 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.983313 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.983331 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.983360 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:55 crc kubenswrapper[5025]: I1004 10:35:55.983379 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:55Z","lastTransitionTime":"2025-10-04T10:35:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.008222 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/3.log" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.014533 5025 scope.go:117] "RemoveContainer" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.015064 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.037113 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.058178 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.076099 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.086553 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.086884 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.087191 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.087362 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.087489 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:56Z","lastTransitionTime":"2025-10-04T10:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.092074 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.111103 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.130697 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.153211 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:40Z\\\",\\\"message\\\":\\\"2025-10-04T10:34:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413\\\\n2025-10-04T10:34:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413 to /host/opt/cni/bin/\\\\n2025-10-04T10:34:55Z [verbose] multus-daemon started\\\\n2025-10-04T10:34:55Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:35:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.171601 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.191269 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.191345 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.191368 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.191402 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.191428 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:56Z","lastTransitionTime":"2025-10-04T10:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.204619 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:54Z\\\",\\\"message\\\":\\\"004 10:35:54.376544 7061 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 10:35:54.376630 7061 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 10:35:54.376664 7061 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 10:35:54.376847 7061 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 10:35:54.376897 7061 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 10:35:54.376905 7061 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 10:35:54.376916 7061 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 10:35:54.376922 7061 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 10:35:54.376946 7061 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 10:35:54.376963 7061 factory.go:656] Stopping watch factory\\\\nI1004 10:35:54.376974 7061 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 10:35:54.376979 7061 ovnkube.go:599] Stopped ovnkube\\\\nI1004 10:35:54.377002 7061 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 10:35:54.377008 7061 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 10:35:54.377036 7061 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 10:35:54.377030 7061 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.224415 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.228830 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.229068 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.229138 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.229179 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.229142828 +0000 UTC m=+148.654109778 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.229232 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.229402 5025 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.229494 5025 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.229524 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.229582 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.229611 5025 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.229550 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.22951328 +0000 UTC m=+148.654480380 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.230389 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.229702437 +0000 UTC m=+148.654669327 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.230526 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.230438851 +0000 UTC m=+148.655405771 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.252507 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.267148 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.279859 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.294695 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.294757 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.294780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.294810 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.294834 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:56Z","lastTransitionTime":"2025-10-04T10:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.295831 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a4bf1ab-6621-4e87-a088-139aaa7c8406\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf053ec8fd21d01980e6f7f6d5f24328ff39cf7d641f4ed959ce0ca034f4fbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f018adfdcab8bd60cd03879d0807145bf0e8b398a9a138739387f0f420c7989b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6bc1b4dcad8484ee9c6f82153ed8d93448d8c3540d44e6990981983f9573cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.309953 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.326477 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.330722 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.331100 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.331186 5025 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.331218 5025 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.331382 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.331357926 +0000 UTC m=+148.756324846 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.347936 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.361910 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:56Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.397834 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.397883 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.397901 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.397927 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.397946 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:56Z","lastTransitionTime":"2025-10-04T10:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.410420 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.410474 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.410593 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.410620 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.410803 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:56 crc kubenswrapper[5025]: E1004 10:35:56.410939 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.500630 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.500696 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.500712 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.500737 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.500755 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:56Z","lastTransitionTime":"2025-10-04T10:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.603892 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.603969 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.603993 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.604068 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.604097 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:56Z","lastTransitionTime":"2025-10-04T10:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.707614 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.707680 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.707701 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.707726 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.707746 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:56Z","lastTransitionTime":"2025-10-04T10:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.810578 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.810627 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.810639 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.810660 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.810677 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:56Z","lastTransitionTime":"2025-10-04T10:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.914919 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.915001 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.915079 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.915114 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:56 crc kubenswrapper[5025]: I1004 10:35:56.915139 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:56Z","lastTransitionTime":"2025-10-04T10:35:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.017599 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.017652 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.017669 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.017694 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.017714 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:57Z","lastTransitionTime":"2025-10-04T10:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.120716 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.120800 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.120824 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.120853 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.120872 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:57Z","lastTransitionTime":"2025-10-04T10:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.224304 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.224403 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.224437 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.224470 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.224494 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:57Z","lastTransitionTime":"2025-10-04T10:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.328403 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.328522 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.328542 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.328566 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.328584 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:57Z","lastTransitionTime":"2025-10-04T10:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.411183 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:57 crc kubenswrapper[5025]: E1004 10:35:57.411354 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.431356 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.431438 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.431468 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.431504 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.431531 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:57Z","lastTransitionTime":"2025-10-04T10:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.535276 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.535318 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.535329 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.535372 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.535387 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:57Z","lastTransitionTime":"2025-10-04T10:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.639338 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.639390 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.639399 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.639415 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.639428 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:57Z","lastTransitionTime":"2025-10-04T10:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.742410 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.742450 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.742464 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.742480 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.742489 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:57Z","lastTransitionTime":"2025-10-04T10:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.846360 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.846416 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.846431 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.846451 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.846465 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:57Z","lastTransitionTime":"2025-10-04T10:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.949489 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.949557 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.949574 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.949592 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:57 crc kubenswrapper[5025]: I1004 10:35:57.949606 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:57Z","lastTransitionTime":"2025-10-04T10:35:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.052093 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.052155 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.052172 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.052197 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.052215 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:58Z","lastTransitionTime":"2025-10-04T10:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.155499 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.155584 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.155611 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.155644 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.155668 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:58Z","lastTransitionTime":"2025-10-04T10:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.258730 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.258777 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.258794 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.258826 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.258846 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:58Z","lastTransitionTime":"2025-10-04T10:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.361828 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.361884 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.361900 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.361928 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.361946 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:58Z","lastTransitionTime":"2025-10-04T10:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.411192 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.411241 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:35:58 crc kubenswrapper[5025]: E1004 10:35:58.411392 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:35:58 crc kubenswrapper[5025]: E1004 10:35:58.411595 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.411672 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:35:58 crc kubenswrapper[5025]: E1004 10:35:58.411860 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.465236 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.465404 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.465667 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.466076 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.466114 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:58Z","lastTransitionTime":"2025-10-04T10:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.570097 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.570138 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.570150 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.570166 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.570177 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:58Z","lastTransitionTime":"2025-10-04T10:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.673899 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.673969 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.673994 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.674074 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.674105 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:58Z","lastTransitionTime":"2025-10-04T10:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.776954 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.776997 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.777006 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.777123 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.777141 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:58Z","lastTransitionTime":"2025-10-04T10:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.880074 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.880145 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.880170 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.880199 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.880220 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:58Z","lastTransitionTime":"2025-10-04T10:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.983719 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.983764 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.983776 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.983792 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:58 crc kubenswrapper[5025]: I1004 10:35:58.983805 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:58Z","lastTransitionTime":"2025-10-04T10:35:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.087694 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.087743 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.087752 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.087771 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.087782 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.192463 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.192528 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.192541 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.192563 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.192578 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.296060 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.296113 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.296130 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.296155 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.296175 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.399052 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.399125 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.399148 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.399180 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.399207 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.410444 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:35:59 crc kubenswrapper[5025]: E1004 10:35:59.410632 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.503326 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.503398 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.503424 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.503458 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.503483 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.597474 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.597580 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.597608 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.597639 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.597661 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: E1004 10:35:59.620389 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.626760 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.627232 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.627261 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.627286 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.627304 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: E1004 10:35:59.652075 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.659080 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.659154 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.659184 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.659216 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.659242 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: E1004 10:35:59.681642 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.686379 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.686417 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.686429 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.686446 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.686460 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: E1004 10:35:59.705943 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.711284 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.711356 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.711381 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.711411 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.711436 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: E1004 10:35:59.731605 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:35:59Z is after 2025-08-24T17:21:41Z" Oct 04 10:35:59 crc kubenswrapper[5025]: E1004 10:35:59.731863 5025 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.733713 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.733746 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.733758 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.733777 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.733791 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.836884 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.836965 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.836984 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.837048 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.837068 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.940173 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.940243 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.940261 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.940286 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:35:59 crc kubenswrapper[5025]: I1004 10:35:59.940304 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:35:59Z","lastTransitionTime":"2025-10-04T10:35:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.043529 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.043588 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.043605 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.043637 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.043656 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:00Z","lastTransitionTime":"2025-10-04T10:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.146848 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.146905 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.146921 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.146953 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.146971 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:00Z","lastTransitionTime":"2025-10-04T10:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.249738 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.249790 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.249827 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.249850 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.249866 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:00Z","lastTransitionTime":"2025-10-04T10:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.353206 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.353291 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.353310 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.353338 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.353357 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:00Z","lastTransitionTime":"2025-10-04T10:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.411031 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:00 crc kubenswrapper[5025]: E1004 10:36:00.411171 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.411332 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.411308 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:00 crc kubenswrapper[5025]: E1004 10:36:00.411591 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:00 crc kubenswrapper[5025]: E1004 10:36:00.411705 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.457586 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.457651 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.457669 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.457692 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.457711 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:00Z","lastTransitionTime":"2025-10-04T10:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.560683 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.560752 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.560776 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.560817 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.560835 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:00Z","lastTransitionTime":"2025-10-04T10:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.664338 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.664406 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.664423 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.664449 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.664466 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:00Z","lastTransitionTime":"2025-10-04T10:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.768153 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.768232 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.768255 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.768290 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.768311 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:00Z","lastTransitionTime":"2025-10-04T10:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.872254 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.872310 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.872321 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.872339 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.872356 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:00Z","lastTransitionTime":"2025-10-04T10:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.975251 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.975295 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.975309 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.975328 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:00 crc kubenswrapper[5025]: I1004 10:36:00.975341 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:00Z","lastTransitionTime":"2025-10-04T10:36:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.079040 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.079148 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.079169 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.079195 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.079215 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:01Z","lastTransitionTime":"2025-10-04T10:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.183093 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.183174 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.183198 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.183228 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.183253 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:01Z","lastTransitionTime":"2025-10-04T10:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.286703 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.286771 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.286796 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.286827 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.286844 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:01Z","lastTransitionTime":"2025-10-04T10:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.389530 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.389583 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.389600 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.389624 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.389641 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:01Z","lastTransitionTime":"2025-10-04T10:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.411318 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:01 crc kubenswrapper[5025]: E1004 10:36:01.411550 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.428441 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.492554 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.492616 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.492640 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.492669 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.492693 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:01Z","lastTransitionTime":"2025-10-04T10:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.596556 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.596620 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.596639 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.596663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.596684 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:01Z","lastTransitionTime":"2025-10-04T10:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.705551 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.706302 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.706329 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.706367 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.706390 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:01Z","lastTransitionTime":"2025-10-04T10:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.809963 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.810073 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.810100 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.810132 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.810153 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:01Z","lastTransitionTime":"2025-10-04T10:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.913391 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.913467 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.913489 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.913521 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:01 crc kubenswrapper[5025]: I1004 10:36:01.913541 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:01Z","lastTransitionTime":"2025-10-04T10:36:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.017115 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.017169 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.017178 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.017197 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.017208 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:02Z","lastTransitionTime":"2025-10-04T10:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.120914 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.120968 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.120980 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.120998 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.121034 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:02Z","lastTransitionTime":"2025-10-04T10:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.224688 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.224756 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.224778 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.224805 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.224826 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:02Z","lastTransitionTime":"2025-10-04T10:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.328359 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.328442 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.328476 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.328507 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.328529 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:02Z","lastTransitionTime":"2025-10-04T10:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.410613 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.410661 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:02 crc kubenswrapper[5025]: E1004 10:36:02.410828 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.410620 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:02 crc kubenswrapper[5025]: E1004 10:36:02.411142 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:02 crc kubenswrapper[5025]: E1004 10:36:02.411300 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.431293 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.431360 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.431376 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.431401 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.431420 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:02Z","lastTransitionTime":"2025-10-04T10:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.433186 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.449705 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8fa40ed34896501fdc4a7d7418f8e3e084f3b345f22eccfdae6e794b3bb2d8a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.467239 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-wfm6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"22480b6b-5db5-4368-b3f5-94305571cbff\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a884699d1a9f0d44e26fc35ba2aab5afea26c67f3d867dd7ffae1d0c0a816433\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vs86s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:58Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-wfm6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.482901 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b1edc0-3621-4e40-bc7c-4b8ca407acbb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe4f587a68c86b165a4c59cd17054fb6be4042cebc58056097b1c690112d8b1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aeb6fa2c060c0ab03a0a42488bdd520182b0576029f00cf4b045cea15f891c11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://248f45e8c75985499abe2c0ac98f0db980e2318ceb6dfed2180e51a38121bc92\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a697495d143e947d3255c2e5265370e88dde2e79616e74903799e6aa98fa8fd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.504768 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://45d5dc037bbac5d5eba6774d5babfbb93ab5b0f8af7caa4f897268231c47746e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a20d31fba5e32cb244de9934b409479594b21084293bbcdb29f284131140529c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.527744 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-n49fv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d214492-7532-4f72-b032-b3cc1ae19473\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:40Z\\\",\\\"message\\\":\\\"2025-10-04T10:34:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413\\\\n2025-10-04T10:34:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_d80523ce-b8eb-4f4a-a56f-20ce9a06c413 to /host/opt/cni/bin/\\\\n2025-10-04T10:34:55Z [verbose] multus-daemon started\\\\n2025-10-04T10:34:55Z [verbose] Readiness Indicator file check\\\\n2025-10-04T10:35:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95hf8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-n49fv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.535935 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.535993 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.536038 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.536065 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.536084 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:02Z","lastTransitionTime":"2025-10-04T10:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.549141 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"54919b0d-887d-4727-adfc-e48a66e680ba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5efcab941b7e83e147ec9e96083a4e55f03d2177be1091c992aae7d1e9a7e86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7zdvw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-2dll9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.579829 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"607a1d66-62fc-4dba-9c44-6798f087fb5c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-04T10:35:54Z\\\",\\\"message\\\":\\\"004 10:35:54.376544 7061 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1004 10:35:54.376630 7061 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1004 10:35:54.376664 7061 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1004 10:35:54.376847 7061 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1004 10:35:54.376897 7061 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1004 10:35:54.376905 7061 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1004 10:35:54.376916 7061 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1004 10:35:54.376922 7061 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1004 10:35:54.376946 7061 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1004 10:35:54.376963 7061 factory.go:656] Stopping watch factory\\\\nI1004 10:35:54.376974 7061 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1004 10:35:54.376979 7061 ovnkube.go:599] Stopped ovnkube\\\\nI1004 10:35:54.377002 7061 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1004 10:35:54.377008 7061 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1004 10:35:54.377036 7061 handler.go:208] Removed *v1.Node event handler 2\\\\nI1004 10:35:54.377030 7061 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:35:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z75f4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bsdl9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.595993 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-frc27" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86312d38-63d0-409a-98d4-727f0fb47929\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5hvf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-frc27\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.616787 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://531903abc56e7eb63e32fa3eac29918e1f4e907bdfb89033567d194c86e4d494\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.635877 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.639872 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.639925 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.639941 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.639963 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.639979 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:02Z","lastTransitionTime":"2025-10-04T10:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.651566 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-h9kd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"252f30f4-ff2f-45a1-abbf-da9365c09581\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f089177fa8ec7ded469b99358000b240125c3f1552b613bc72854addd0871595\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c68r2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-h9kd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.683789 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"35b0afab-5f30-48fc-8af8-8dfd73449c0e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fe959459a85adf1ad340e0acdccf169838fb3b34102b4afe04a76e5fcb976d78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2b1d7689121db3ed2954eadb310d6c505ef8a21ed4e425b68ff761cff85aa462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fdaea481699a14215959a19eef509030513c2ce8d60103ef5b4d2a2f637dcaa4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9f86fe7dda6b70accda87ebf9c59a7438ecb51d93574bd9d136d77dcd7ad32f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://56fd14fe72f295f2c3275d6ce5bc9a8089feb4e4d750aec15b4b954af4609274\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb78bcfc620105f2eb6d8d1cbf695f6e57955583e8c453ac4449cbf9356bccd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://759064736f40fa0c0893faeee2b476f9c79693552f90a874a32c350d40bdd41c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2778846d83222b829b5f54ffdf8ce39620759e93485e3f502d57d3716d885409\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.698900 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caa3abf8-7e53-47db-8dcb-11d67a550039\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01b013405faf0a0ebcab7a507811aaf62fdcec773d83acaf20087fe3a5a58f68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52644454e1e9f7ec8fb33d08693d42116674144b05f0ce277a92ec07ffa51e5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52644454e1e9f7ec8fb33d08693d42116674144b05f0ce277a92ec07ffa51e5d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.719077 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d305a083-d3d7-4a9c-8ef2-554d17a5dc84\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73fb68af8f14c24a35237bb864e30c4f313754d5dd205518e0696aa25c08b3d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae9a008b868d3b37b98e6027ef18926cdea6505aa582fd928e62926b8e9d0b71\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed0704e17377c6a450550c020f098dc5b4215407bbad66a4b5b123642634bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5efbe8e932d3a384ed9d0f64496741f922afb46f90ca3e5cb829a235597a2979\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bf323424c1ef3413e2de94ecf4cadd03facc36f7b9255b6a95962f87d5f16bc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-04T10:34:51Z\\\",\\\"message\\\":\\\"Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510857 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1004 10:34:51.510863 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1004 10:34:51.510870 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1004 10:34:51.510873 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1004 10:34:51.510876 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1004 10:34:51.511044 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nI1004 10:34:51.523201 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1759574075\\\\\\\\\\\\\\\" (2025-10-04 10:34:34 +0000 UTC to 2025-11-03 10:34:35 +0000 UTC (now=2025-10-04 10:34:51.523145874 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523416 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1759574076\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1759574075\\\\\\\\\\\\\\\" (2025-10-04 09:34:35 +0000 UTC to 2026-10-04 09:34:35 +0000 UTC (now=2025-10-04 10:34:51.523391361 +0000 UTC))\\\\\\\"\\\\nI1004 10:34:51.523461 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1004 10:34:51.523490 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1004 10:34:51.523544 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1939682795/tls.crt::/tmp/serving-cert-1939682795/tls.key\\\\\\\"\\\\nF1004 10:34:51.523615 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8ea062da3b7ff5d5b4fa91ad5f92fe0ab28004a92ecafc4d48661c86b6d5b029\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d2c44830f0caee5c0d335e845cdd07b7fe2c517fdc2dc2eb723e3ec1070cf42f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.739099 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.743553 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.743709 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.743809 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.743900 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.743987 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:02Z","lastTransitionTime":"2025-10-04T10:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.763826 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"24e4becb-78de-4912-ac5f-d25ee8710f40\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27646a992d2ac5adf797692cafd8190de54eb3784f2234e4c9f29f0ddff8919b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b178b8c35a6ab763984ce5faca56fddcfc0c92b7610d12c7ee69b3975cd428\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:54Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6f9ef6a5979353beae42adca566d982bb8b4980debb9ad6e982026e3d6677fe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://228364d98453272a32d4318f0074bd6a1193e417719008ccfadd5ca07d98189f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://639bc7846372c29bec82bab58f6ebb4cfb9a78a280989f73e562ed9ea90d2cf3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea480f2f1f3b337bca3cd9e3c8f38c7e657f01a726dfe4cd87adf75c53a0082\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e618d24b572cbcd0ae7b054ed5c696e72bd6c3ec769b8b7704bdceb96d01837\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fkmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:53Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-h8qkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.779085 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"44d57967-21de-4a17-9fe3-b5a7678e7a82\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e53547d21ce1c98d456b4f577de76b12bbfd931d6261ebff55005dae13928e32\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://68dbd460f94fd4ca26a712e2a2d91c2a466533e27513fc95ef29291a534033b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:35:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bdpgt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:35:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-7hlrk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.793874 5025 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5a4bf1ab-6621-4e87-a088-139aaa7c8406\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:35:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-04T10:34:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cf053ec8fd21d01980e6f7f6d5f24328ff39cf7d641f4ed959ce0ca034f4fbbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f018adfdcab8bd60cd03879d0807145bf0e8b398a9a138739387f0f420c7989b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b6bc1b4dcad8484ee9c6f82153ed8d93448d8c3540d44e6990981983f9573cbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-04T10:34:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c20eee57d2d7453ae01bdc1f642f3e73c3b9871c7084e3ac824f953d3e95e42\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-04T10:34:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-04T10:34:33Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-04T10:34:32Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:02Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.846812 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.846889 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.846915 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.846946 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.846965 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:02Z","lastTransitionTime":"2025-10-04T10:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.950686 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.950763 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.950783 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.950808 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:02 crc kubenswrapper[5025]: I1004 10:36:02.950826 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:02Z","lastTransitionTime":"2025-10-04T10:36:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.053122 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.053188 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.053208 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.053237 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.053260 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:03Z","lastTransitionTime":"2025-10-04T10:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.156063 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.156107 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.156118 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.156135 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.156145 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:03Z","lastTransitionTime":"2025-10-04T10:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.259292 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.259356 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.259372 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.259401 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.259420 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:03Z","lastTransitionTime":"2025-10-04T10:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.363402 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.363834 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.363986 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.364185 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.364326 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:03Z","lastTransitionTime":"2025-10-04T10:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.410402 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:03 crc kubenswrapper[5025]: E1004 10:36:03.410778 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.477529 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.477573 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.477590 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.477625 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.477643 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:03Z","lastTransitionTime":"2025-10-04T10:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.580616 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.580666 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.580679 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.580694 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.580704 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:03Z","lastTransitionTime":"2025-10-04T10:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.683926 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.683989 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.684006 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.684061 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.684080 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:03Z","lastTransitionTime":"2025-10-04T10:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.786277 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.786565 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.786659 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.786751 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.786835 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:03Z","lastTransitionTime":"2025-10-04T10:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.889993 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.890069 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.890085 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.890110 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.890126 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:03Z","lastTransitionTime":"2025-10-04T10:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.993195 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.993296 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.993316 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.993341 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:03 crc kubenswrapper[5025]: I1004 10:36:03.993360 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:03Z","lastTransitionTime":"2025-10-04T10:36:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.096184 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.096262 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.096281 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.096306 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.096325 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:04Z","lastTransitionTime":"2025-10-04T10:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.198649 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.198717 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.198735 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.198761 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.198780 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:04Z","lastTransitionTime":"2025-10-04T10:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.301958 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.302067 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.302094 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.302123 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.302147 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:04Z","lastTransitionTime":"2025-10-04T10:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.405805 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.405897 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.405915 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.405938 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.405956 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:04Z","lastTransitionTime":"2025-10-04T10:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.416116 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:04 crc kubenswrapper[5025]: E1004 10:36:04.416230 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.416325 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.416366 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:04 crc kubenswrapper[5025]: E1004 10:36:04.416457 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:04 crc kubenswrapper[5025]: E1004 10:36:04.416606 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.509316 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.509376 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.509393 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.509416 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.509436 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:04Z","lastTransitionTime":"2025-10-04T10:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.612766 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.612825 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.612842 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.612869 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.612887 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:04Z","lastTransitionTime":"2025-10-04T10:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.715960 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.716001 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.716027 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.716043 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.716069 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:04Z","lastTransitionTime":"2025-10-04T10:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.819361 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.819405 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.819417 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.819432 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.819442 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:04Z","lastTransitionTime":"2025-10-04T10:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.922813 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.922860 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.922874 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.922893 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:04 crc kubenswrapper[5025]: I1004 10:36:04.922907 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:04Z","lastTransitionTime":"2025-10-04T10:36:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.026565 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.026644 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.026663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.026692 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.026712 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:05Z","lastTransitionTime":"2025-10-04T10:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.130648 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.130727 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.130746 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.130773 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.130793 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:05Z","lastTransitionTime":"2025-10-04T10:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.233795 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.233852 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.233866 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.233887 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.233902 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:05Z","lastTransitionTime":"2025-10-04T10:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.337428 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.337533 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.337561 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.337595 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.337621 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:05Z","lastTransitionTime":"2025-10-04T10:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.411133 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:05 crc kubenswrapper[5025]: E1004 10:36:05.411301 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.440284 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.440330 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.440342 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.440361 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.440373 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:05Z","lastTransitionTime":"2025-10-04T10:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.543094 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.543151 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.543169 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.543191 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.543206 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:05Z","lastTransitionTime":"2025-10-04T10:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.645541 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.645721 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.645826 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.645878 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.645902 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:05Z","lastTransitionTime":"2025-10-04T10:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.749612 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.749680 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.749704 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.749729 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.749748 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:05Z","lastTransitionTime":"2025-10-04T10:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.853816 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.853893 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.853919 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.853949 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.853972 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:05Z","lastTransitionTime":"2025-10-04T10:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.956966 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.957142 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.957170 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.957235 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:05 crc kubenswrapper[5025]: I1004 10:36:05.957257 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:05Z","lastTransitionTime":"2025-10-04T10:36:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.061045 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.061109 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.061128 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.061154 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.061173 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:06Z","lastTransitionTime":"2025-10-04T10:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.164575 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.164660 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.164682 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.164708 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.164726 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:06Z","lastTransitionTime":"2025-10-04T10:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.267970 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.268064 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.268084 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.268110 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.268128 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:06Z","lastTransitionTime":"2025-10-04T10:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.370876 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.370933 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.370951 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.370973 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.370993 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:06Z","lastTransitionTime":"2025-10-04T10:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.411084 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.411212 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:06 crc kubenswrapper[5025]: E1004 10:36:06.411404 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.411639 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:06 crc kubenswrapper[5025]: E1004 10:36:06.411783 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:06 crc kubenswrapper[5025]: E1004 10:36:06.412089 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.474195 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.474242 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.474260 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.474281 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.474298 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:06Z","lastTransitionTime":"2025-10-04T10:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.580046 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.580134 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.580158 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.580193 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.580217 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:06Z","lastTransitionTime":"2025-10-04T10:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.683926 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.684002 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.684050 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.684076 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.684096 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:06Z","lastTransitionTime":"2025-10-04T10:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.787832 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.787905 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.787923 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.787950 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.787973 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:06Z","lastTransitionTime":"2025-10-04T10:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.891069 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.891143 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.891161 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.891185 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.891203 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:06Z","lastTransitionTime":"2025-10-04T10:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.994115 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.994183 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.994202 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.994227 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:06 crc kubenswrapper[5025]: I1004 10:36:06.994246 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:06Z","lastTransitionTime":"2025-10-04T10:36:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.097498 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.097572 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.097597 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.097628 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.097649 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:07Z","lastTransitionTime":"2025-10-04T10:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.200503 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.200562 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.200579 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.200602 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.200620 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:07Z","lastTransitionTime":"2025-10-04T10:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.303501 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.303591 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.303619 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.303651 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.303672 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:07Z","lastTransitionTime":"2025-10-04T10:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.407488 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.407598 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.407620 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.407648 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.407671 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:07Z","lastTransitionTime":"2025-10-04T10:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.410842 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:07 crc kubenswrapper[5025]: E1004 10:36:07.411051 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.510507 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.510570 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.510588 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.510612 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.510632 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:07Z","lastTransitionTime":"2025-10-04T10:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.613271 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.613335 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.613353 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.613382 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.613401 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:07Z","lastTransitionTime":"2025-10-04T10:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.716066 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.716138 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.716161 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.716189 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.716210 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:07Z","lastTransitionTime":"2025-10-04T10:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.819336 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.819386 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.819399 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.819419 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.819432 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:07Z","lastTransitionTime":"2025-10-04T10:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.922973 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.923097 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.923121 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.923149 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:07 crc kubenswrapper[5025]: I1004 10:36:07.923169 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:07Z","lastTransitionTime":"2025-10-04T10:36:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.025730 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.025811 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.025853 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.025876 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.025890 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:08Z","lastTransitionTime":"2025-10-04T10:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.129274 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.129329 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.129338 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.129357 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.129375 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:08Z","lastTransitionTime":"2025-10-04T10:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.233372 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.233439 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.233460 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.233488 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.233505 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:08Z","lastTransitionTime":"2025-10-04T10:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.337297 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.337347 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.337359 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.337377 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.337390 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:08Z","lastTransitionTime":"2025-10-04T10:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.411191 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.411295 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.411312 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:08 crc kubenswrapper[5025]: E1004 10:36:08.411484 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:08 crc kubenswrapper[5025]: E1004 10:36:08.411680 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:08 crc kubenswrapper[5025]: E1004 10:36:08.411871 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.439943 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.440006 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.440064 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.440095 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.440116 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:08Z","lastTransitionTime":"2025-10-04T10:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.543926 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.544001 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.544039 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.544053 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.544063 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:08Z","lastTransitionTime":"2025-10-04T10:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.647644 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.647723 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.647748 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.647781 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.647805 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:08Z","lastTransitionTime":"2025-10-04T10:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.754287 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.754356 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.754380 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.754405 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.754425 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:08Z","lastTransitionTime":"2025-10-04T10:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.857818 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.857877 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.857894 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.857917 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.857939 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:08Z","lastTransitionTime":"2025-10-04T10:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.961432 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.961584 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.961606 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.961669 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:08 crc kubenswrapper[5025]: I1004 10:36:08.961694 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:08Z","lastTransitionTime":"2025-10-04T10:36:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.064610 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.064670 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.064685 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.064707 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.064728 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.167811 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.167881 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.167898 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.167921 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.167940 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.271392 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.271462 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.271485 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.271514 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.271536 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.374858 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.374923 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.374940 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.374962 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.374979 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.411066 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:09 crc kubenswrapper[5025]: E1004 10:36:09.411303 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.412445 5025 scope.go:117] "RemoveContainer" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:36:09 crc kubenswrapper[5025]: E1004 10:36:09.412702 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.478378 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.478443 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.478461 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.478484 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.478505 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.581535 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.581618 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.581636 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.582160 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.582216 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.685992 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.686075 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.686094 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.686121 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.686139 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.789369 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.789432 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.789453 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.789478 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.789495 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.869589 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.869657 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.869675 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.869697 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.869714 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: E1004 10:36:09.886532 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:09Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.891704 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.891762 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.891779 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.891803 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.891821 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: E1004 10:36:09.911884 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:09Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.916846 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.916905 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.916923 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.916948 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.916968 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: E1004 10:36:09.943434 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:09Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.949626 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.949663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.949674 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.949690 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.949702 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: E1004 10:36:09.966232 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:09Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.971431 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.971465 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.971475 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.971489 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.971501 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:09 crc kubenswrapper[5025]: E1004 10:36:09.989270 5025 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-04T10:36:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"707d7ca1-ef0c-4cef-a8ca-1f574b989b45\\\",\\\"systemUUID\\\":\\\"422640ad-1ab6-4871-9728-16b07446efb8\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-04T10:36:09Z is after 2025-08-24T17:21:41Z" Oct 04 10:36:09 crc kubenswrapper[5025]: E1004 10:36:09.989748 5025 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.991708 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.991783 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.991811 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.991839 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:09 crc kubenswrapper[5025]: I1004 10:36:09.991858 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:09Z","lastTransitionTime":"2025-10-04T10:36:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.094704 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.094771 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.094793 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.094821 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.094843 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:10Z","lastTransitionTime":"2025-10-04T10:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.198427 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.198507 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.198527 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.198611 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.198638 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:10Z","lastTransitionTime":"2025-10-04T10:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.301774 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.301869 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.301893 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.301922 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.301941 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:10Z","lastTransitionTime":"2025-10-04T10:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.405005 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.405115 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.405141 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.405169 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.405190 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:10Z","lastTransitionTime":"2025-10-04T10:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.410493 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.410692 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:10 crc kubenswrapper[5025]: E1004 10:36:10.410691 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.410732 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:10 crc kubenswrapper[5025]: E1004 10:36:10.411059 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:10 crc kubenswrapper[5025]: E1004 10:36:10.411252 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.508420 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.508491 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.508513 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.508540 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.508561 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:10Z","lastTransitionTime":"2025-10-04T10:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.611114 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.611220 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.611238 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.611262 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.611284 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:10Z","lastTransitionTime":"2025-10-04T10:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.714153 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.714223 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.714241 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.714266 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.714286 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:10Z","lastTransitionTime":"2025-10-04T10:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.817796 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.817876 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.817888 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.817920 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.817938 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:10Z","lastTransitionTime":"2025-10-04T10:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.920958 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.921055 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.921075 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.921100 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:10 crc kubenswrapper[5025]: I1004 10:36:10.921119 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:10Z","lastTransitionTime":"2025-10-04T10:36:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.025105 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.025181 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.025198 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.025221 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.025240 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:11Z","lastTransitionTime":"2025-10-04T10:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.128539 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.128609 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.128633 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.128672 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.128697 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:11Z","lastTransitionTime":"2025-10-04T10:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.231154 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.231242 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.231267 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.231787 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.232106 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:11Z","lastTransitionTime":"2025-10-04T10:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.335597 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.335647 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.335663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.335684 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.335699 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:11Z","lastTransitionTime":"2025-10-04T10:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.410811 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:11 crc kubenswrapper[5025]: E1004 10:36:11.411010 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.438150 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.438204 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.438242 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.438261 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.438272 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:11Z","lastTransitionTime":"2025-10-04T10:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.541466 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.541528 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.541545 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.541569 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.541586 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:11Z","lastTransitionTime":"2025-10-04T10:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.610500 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:11 crc kubenswrapper[5025]: E1004 10:36:11.610699 5025 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:36:11 crc kubenswrapper[5025]: E1004 10:36:11.610773 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs podName:86312d38-63d0-409a-98d4-727f0fb47929 nodeName:}" failed. No retries permitted until 2025-10-04 10:37:15.610748556 +0000 UTC m=+164.035715476 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs") pod "network-metrics-daemon-frc27" (UID: "86312d38-63d0-409a-98d4-727f0fb47929") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.644631 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.644679 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.644695 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.644716 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.644733 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:11Z","lastTransitionTime":"2025-10-04T10:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.748414 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.748497 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.748517 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.748542 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.748599 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:11Z","lastTransitionTime":"2025-10-04T10:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.852627 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.852692 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.852709 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.852733 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.852753 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:11Z","lastTransitionTime":"2025-10-04T10:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.956079 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.956157 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.956181 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.956213 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:11 crc kubenswrapper[5025]: I1004 10:36:11.956234 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:11Z","lastTransitionTime":"2025-10-04T10:36:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.059554 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.059638 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.059664 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.059700 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.059773 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:12Z","lastTransitionTime":"2025-10-04T10:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.163215 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.163290 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.163309 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.163336 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.163353 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:12Z","lastTransitionTime":"2025-10-04T10:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.266909 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.266966 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.266982 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.267005 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.267068 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:12Z","lastTransitionTime":"2025-10-04T10:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.369987 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.370121 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.370145 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.370181 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.370203 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:12Z","lastTransitionTime":"2025-10-04T10:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.410796 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.410943 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:12 crc kubenswrapper[5025]: E1004 10:36:12.411257 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.411334 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:12 crc kubenswrapper[5025]: E1004 10:36:12.411471 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:12 crc kubenswrapper[5025]: E1004 10:36:12.411655 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.473513 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.473548 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.473559 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.473575 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.473586 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:12Z","lastTransitionTime":"2025-10-04T10:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.507319 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=79.507281875 podStartE2EDuration="1m19.507281875s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.478884399 +0000 UTC m=+100.903851299" watchObservedRunningTime="2025-10-04 10:36:12.507281875 +0000 UTC m=+100.932248755" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.547234 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-h9kd5" podStartSLOduration=79.547211267 podStartE2EDuration="1m19.547211267s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.525869025 +0000 UTC m=+100.950835905" watchObservedRunningTime="2025-10-04 10:36:12.547211267 +0000 UTC m=+100.972178147" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.547360 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.547355731 podStartE2EDuration="44.547355731s" podCreationTimestamp="2025-10-04 10:35:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.546354808 +0000 UTC m=+100.971321688" watchObservedRunningTime="2025-10-04 10:36:12.547355731 +0000 UTC m=+100.972322611" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.564858 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=11.564839274 podStartE2EDuration="11.564839274s" podCreationTimestamp="2025-10-04 10:36:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.562882479 +0000 UTC m=+100.987849359" watchObservedRunningTime="2025-10-04 10:36:12.564839274 +0000 UTC m=+100.989806154" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.576721 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.576779 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.576795 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.576817 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.576833 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:12Z","lastTransitionTime":"2025-10-04T10:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.590714 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=80.590677306 podStartE2EDuration="1m20.590677306s" podCreationTimestamp="2025-10-04 10:34:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.586422054 +0000 UTC m=+101.011388964" watchObservedRunningTime="2025-10-04 10:36:12.590677306 +0000 UTC m=+101.015644216" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.629303 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-h8qkk" podStartSLOduration=79.629277853 podStartE2EDuration="1m19.629277853s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.628841958 +0000 UTC m=+101.053808848" watchObservedRunningTime="2025-10-04 10:36:12.629277853 +0000 UTC m=+101.054244773" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.645995 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-7hlrk" podStartSLOduration=78.645964279 podStartE2EDuration="1m18.645964279s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.645322038 +0000 UTC m=+101.070288928" watchObservedRunningTime="2025-10-04 10:36:12.645964279 +0000 UTC m=+101.070931199" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.679279 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.679351 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.679371 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.679398 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.679418 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:12Z","lastTransitionTime":"2025-10-04T10:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.688110 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=78.688079493 podStartE2EDuration="1m18.688079493s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.666478103 +0000 UTC m=+101.091444983" watchObservedRunningTime="2025-10-04 10:36:12.688079493 +0000 UTC m=+101.113046373" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.715617 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-wfm6t" podStartSLOduration=78.71559612 podStartE2EDuration="1m18.71559612s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.715243319 +0000 UTC m=+101.140210229" watchObservedRunningTime="2025-10-04 10:36:12.71559612 +0000 UTC m=+101.140563000" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.782514 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.782926 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.782996 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.783107 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.783185 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:12Z","lastTransitionTime":"2025-10-04T10:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.804627 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-n49fv" podStartSLOduration=79.804596118 podStartE2EDuration="1m19.804596118s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.786833505 +0000 UTC m=+101.211800385" watchObservedRunningTime="2025-10-04 10:36:12.804596118 +0000 UTC m=+101.229562998" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.805861 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podStartSLOduration=79.8058551 podStartE2EDuration="1m19.8058551s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:12.804539606 +0000 UTC m=+101.229506486" watchObservedRunningTime="2025-10-04 10:36:12.8058551 +0000 UTC m=+101.230821980" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.886686 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.886750 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.886760 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.886775 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.886784 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:12Z","lastTransitionTime":"2025-10-04T10:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.989382 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.989439 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.989486 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.989505 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:12 crc kubenswrapper[5025]: I1004 10:36:12.989517 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:12Z","lastTransitionTime":"2025-10-04T10:36:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.092762 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.092830 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.092849 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.092881 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.092903 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:13Z","lastTransitionTime":"2025-10-04T10:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.196261 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.196330 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.196340 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.196391 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.196407 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:13Z","lastTransitionTime":"2025-10-04T10:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.299305 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.299406 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.299432 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.299472 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.299514 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:13Z","lastTransitionTime":"2025-10-04T10:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.402958 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.403149 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.403170 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.403196 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.403214 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:13Z","lastTransitionTime":"2025-10-04T10:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.410429 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:13 crc kubenswrapper[5025]: E1004 10:36:13.410612 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.506104 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.506185 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.506211 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.506242 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.506263 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:13Z","lastTransitionTime":"2025-10-04T10:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.609842 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.609915 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.609938 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.609968 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.609987 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:13Z","lastTransitionTime":"2025-10-04T10:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.713505 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.713588 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.713608 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.713639 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.713658 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:13Z","lastTransitionTime":"2025-10-04T10:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.816837 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.816990 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.817053 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.817087 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.817112 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:13Z","lastTransitionTime":"2025-10-04T10:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.920194 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.920257 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.920273 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.920295 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:13 crc kubenswrapper[5025]: I1004 10:36:13.920314 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:13Z","lastTransitionTime":"2025-10-04T10:36:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.023729 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.023811 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.023829 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.023857 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.023876 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:14Z","lastTransitionTime":"2025-10-04T10:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.127158 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.127233 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.127250 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.127274 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.127292 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:14Z","lastTransitionTime":"2025-10-04T10:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.229327 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.229382 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.229398 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.229420 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.229433 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:14Z","lastTransitionTime":"2025-10-04T10:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.332543 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.332601 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.332613 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.332638 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.332652 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:14Z","lastTransitionTime":"2025-10-04T10:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.410575 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.410587 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.410758 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:14 crc kubenswrapper[5025]: E1004 10:36:14.410871 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:14 crc kubenswrapper[5025]: E1004 10:36:14.411119 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:14 crc kubenswrapper[5025]: E1004 10:36:14.411351 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.435696 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.435740 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.435752 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.435770 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.435785 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:14Z","lastTransitionTime":"2025-10-04T10:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.537980 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.538090 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.538112 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.538179 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.538211 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:14Z","lastTransitionTime":"2025-10-04T10:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.643619 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.643676 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.643693 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.643717 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.643736 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:14Z","lastTransitionTime":"2025-10-04T10:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.747472 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.747542 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.747555 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.747579 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.747597 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:14Z","lastTransitionTime":"2025-10-04T10:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.850904 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.850988 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.851003 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.851053 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.851072 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:14Z","lastTransitionTime":"2025-10-04T10:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.953597 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.953643 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.953652 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.953666 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:14 crc kubenswrapper[5025]: I1004 10:36:14.953677 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:14Z","lastTransitionTime":"2025-10-04T10:36:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.056480 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.056572 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.056594 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.056628 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.056645 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:15Z","lastTransitionTime":"2025-10-04T10:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.159664 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.159716 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.159728 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.159746 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.159759 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:15Z","lastTransitionTime":"2025-10-04T10:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.262881 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.262951 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.262970 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.262993 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.263005 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:15Z","lastTransitionTime":"2025-10-04T10:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.365259 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.365311 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.365324 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.365341 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.365353 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:15Z","lastTransitionTime":"2025-10-04T10:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.411248 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:15 crc kubenswrapper[5025]: E1004 10:36:15.411432 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.467282 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.467343 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.467354 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.467374 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.467386 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:15Z","lastTransitionTime":"2025-10-04T10:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.570257 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.570327 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.570344 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.570367 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.570384 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:15Z","lastTransitionTime":"2025-10-04T10:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.673768 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.673851 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.673870 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.673893 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.673912 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:15Z","lastTransitionTime":"2025-10-04T10:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.776482 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.776523 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.776533 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.776546 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.776558 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:15Z","lastTransitionTime":"2025-10-04T10:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.878939 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.879001 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.879037 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.879059 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.879077 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:15Z","lastTransitionTime":"2025-10-04T10:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.981393 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.981442 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.981456 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.981480 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:15 crc kubenswrapper[5025]: I1004 10:36:15.981495 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:15Z","lastTransitionTime":"2025-10-04T10:36:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.084280 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.084322 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.084330 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.084345 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.084353 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:16Z","lastTransitionTime":"2025-10-04T10:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.187292 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.187353 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.187371 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.187393 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.187410 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:16Z","lastTransitionTime":"2025-10-04T10:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.290950 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.291100 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.291116 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.291139 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.291157 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:16Z","lastTransitionTime":"2025-10-04T10:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.394663 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.394777 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.394829 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.394852 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.394868 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:16Z","lastTransitionTime":"2025-10-04T10:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.410204 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.410295 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:16 crc kubenswrapper[5025]: E1004 10:36:16.410407 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.410448 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:16 crc kubenswrapper[5025]: E1004 10:36:16.410540 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:16 crc kubenswrapper[5025]: E1004 10:36:16.410669 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.497761 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.497834 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.497859 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.497890 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.497912 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:16Z","lastTransitionTime":"2025-10-04T10:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.601054 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.601109 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.601125 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.601142 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.601156 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:16Z","lastTransitionTime":"2025-10-04T10:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.704105 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.704150 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.704162 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.704178 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.704190 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:16Z","lastTransitionTime":"2025-10-04T10:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.807339 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.807387 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.807398 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.807414 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.807425 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:16Z","lastTransitionTime":"2025-10-04T10:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.910078 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.910135 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.910149 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.910169 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:16 crc kubenswrapper[5025]: I1004 10:36:16.910183 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:16Z","lastTransitionTime":"2025-10-04T10:36:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.012233 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.012281 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.012292 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.012310 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.012323 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:17Z","lastTransitionTime":"2025-10-04T10:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.114496 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.114559 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.114571 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.114590 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.114601 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:17Z","lastTransitionTime":"2025-10-04T10:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.217621 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.217668 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.217681 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.217695 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.217707 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:17Z","lastTransitionTime":"2025-10-04T10:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.326316 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.326362 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.326370 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.326382 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.326391 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:17Z","lastTransitionTime":"2025-10-04T10:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.410804 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:17 crc kubenswrapper[5025]: E1004 10:36:17.411179 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.428421 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.428465 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.428476 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.428491 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.428503 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:17Z","lastTransitionTime":"2025-10-04T10:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.531220 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.531268 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.531278 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.531291 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.531301 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:17Z","lastTransitionTime":"2025-10-04T10:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.634275 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.634315 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.634326 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.634341 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.634351 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:17Z","lastTransitionTime":"2025-10-04T10:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.737052 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.737114 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.737125 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.737143 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.737158 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:17Z","lastTransitionTime":"2025-10-04T10:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.840426 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.840491 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.840530 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.840550 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.840600 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:17Z","lastTransitionTime":"2025-10-04T10:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.942949 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.942999 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.943041 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.943059 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:17 crc kubenswrapper[5025]: I1004 10:36:17.943070 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:17Z","lastTransitionTime":"2025-10-04T10:36:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.046384 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.046462 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.046485 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.046512 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.046531 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:18Z","lastTransitionTime":"2025-10-04T10:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.148579 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.148658 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.148675 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.148698 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.149103 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:18Z","lastTransitionTime":"2025-10-04T10:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.252465 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.252529 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.252548 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.252572 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.252589 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:18Z","lastTransitionTime":"2025-10-04T10:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.355039 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.355101 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.355119 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.355142 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.355159 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:18Z","lastTransitionTime":"2025-10-04T10:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.411261 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.411302 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:18 crc kubenswrapper[5025]: E1004 10:36:18.411461 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.411494 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:18 crc kubenswrapper[5025]: E1004 10:36:18.411648 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:18 crc kubenswrapper[5025]: E1004 10:36:18.411752 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.458341 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.458396 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.458418 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.458446 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.458468 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:18Z","lastTransitionTime":"2025-10-04T10:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.561349 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.561422 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.561449 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.561478 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.561502 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:18Z","lastTransitionTime":"2025-10-04T10:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.664603 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.664676 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.664696 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.664721 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.664739 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:18Z","lastTransitionTime":"2025-10-04T10:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.767324 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.767389 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.767420 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.767445 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.767462 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:18Z","lastTransitionTime":"2025-10-04T10:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.870851 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.870911 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.870928 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.870952 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.870969 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:18Z","lastTransitionTime":"2025-10-04T10:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.973715 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.973768 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.973780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.973797 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:18 crc kubenswrapper[5025]: I1004 10:36:18.973808 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:18Z","lastTransitionTime":"2025-10-04T10:36:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.076110 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.076167 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.076182 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.076205 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.076220 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:19Z","lastTransitionTime":"2025-10-04T10:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.178537 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.178628 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.178664 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.178694 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.178721 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:19Z","lastTransitionTime":"2025-10-04T10:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.281415 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.281478 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.281494 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.281519 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.281536 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:19Z","lastTransitionTime":"2025-10-04T10:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.383883 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.383953 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.383979 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.384007 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.384063 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:19Z","lastTransitionTime":"2025-10-04T10:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.410874 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:19 crc kubenswrapper[5025]: E1004 10:36:19.411068 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.486780 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.486843 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.486861 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.486886 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.486907 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:19Z","lastTransitionTime":"2025-10-04T10:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.590520 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.590615 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.590634 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.590677 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.590689 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:19Z","lastTransitionTime":"2025-10-04T10:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.693348 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.693426 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.693447 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.693473 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.693495 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:19Z","lastTransitionTime":"2025-10-04T10:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.796852 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.796896 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.796907 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.796921 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.796932 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:19Z","lastTransitionTime":"2025-10-04T10:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.899834 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.899898 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.899952 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.899978 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:19 crc kubenswrapper[5025]: I1004 10:36:19.899995 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:19Z","lastTransitionTime":"2025-10-04T10:36:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.003112 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.003175 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.003192 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.003214 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.003231 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:20Z","lastTransitionTime":"2025-10-04T10:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.105787 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.105846 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.105862 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.105883 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.105899 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:20Z","lastTransitionTime":"2025-10-04T10:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.113478 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.113518 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.113532 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.113548 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.113561 5025 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-04T10:36:20Z","lastTransitionTime":"2025-10-04T10:36:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.173505 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l"] Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.173874 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.182321 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.182333 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.189204 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.190166 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.202341 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.202481 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.202683 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.203307 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.206504 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.307432 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.307477 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.307493 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.307523 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.307557 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.307800 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.307867 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.309167 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.314454 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.347948 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4rf6l\" (UID: \"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.411336 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.411405 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.411433 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:20 crc kubenswrapper[5025]: E1004 10:36:20.411573 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:20 crc kubenswrapper[5025]: E1004 10:36:20.411755 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:20 crc kubenswrapper[5025]: E1004 10:36:20.411885 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:20 crc kubenswrapper[5025]: I1004 10:36:20.501184 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" Oct 04 10:36:20 crc kubenswrapper[5025]: W1004 10:36:20.515660 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0ffd7a7_a7c8_403f_8e9a_4f9099f760c4.slice/crio-e1a24d301c5d0e7e28164d7a0699c4e6083630199c7d8c836e040d6f6c09e2fb WatchSource:0}: Error finding container e1a24d301c5d0e7e28164d7a0699c4e6083630199c7d8c836e040d6f6c09e2fb: Status 404 returned error can't find the container with id e1a24d301c5d0e7e28164d7a0699c4e6083630199c7d8c836e040d6f6c09e2fb Oct 04 10:36:21 crc kubenswrapper[5025]: I1004 10:36:21.107608 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" event={"ID":"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4","Type":"ContainerStarted","Data":"20b078d025a3c054c82717736b58ac378aa4ab9f907904acbf35580e27b6b73b"} Oct 04 10:36:21 crc kubenswrapper[5025]: I1004 10:36:21.107665 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" event={"ID":"d0ffd7a7-a7c8-403f-8e9a-4f9099f760c4","Type":"ContainerStarted","Data":"e1a24d301c5d0e7e28164d7a0699c4e6083630199c7d8c836e040d6f6c09e2fb"} Oct 04 10:36:21 crc kubenswrapper[5025]: I1004 10:36:21.124065 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4rf6l" podStartSLOduration=88.124050326 podStartE2EDuration="1m28.124050326s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:21.123610872 +0000 UTC m=+109.548577762" watchObservedRunningTime="2025-10-04 10:36:21.124050326 +0000 UTC m=+109.549017206" Oct 04 10:36:21 crc kubenswrapper[5025]: I1004 10:36:21.410403 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:21 crc kubenswrapper[5025]: E1004 10:36:21.411095 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:21 crc kubenswrapper[5025]: I1004 10:36:21.411593 5025 scope.go:117] "RemoveContainer" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:36:21 crc kubenswrapper[5025]: E1004 10:36:21.411854 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bsdl9_openshift-ovn-kubernetes(607a1d66-62fc-4dba-9c44-6798f087fb5c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" Oct 04 10:36:22 crc kubenswrapper[5025]: I1004 10:36:22.410552 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:22 crc kubenswrapper[5025]: I1004 10:36:22.410656 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:22 crc kubenswrapper[5025]: E1004 10:36:22.411715 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:22 crc kubenswrapper[5025]: I1004 10:36:22.411770 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:22 crc kubenswrapper[5025]: E1004 10:36:22.411930 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:22 crc kubenswrapper[5025]: E1004 10:36:22.412337 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:23 crc kubenswrapper[5025]: I1004 10:36:23.410771 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:23 crc kubenswrapper[5025]: E1004 10:36:23.410876 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:24 crc kubenswrapper[5025]: I1004 10:36:24.411071 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:24 crc kubenswrapper[5025]: I1004 10:36:24.411145 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:24 crc kubenswrapper[5025]: E1004 10:36:24.411416 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:24 crc kubenswrapper[5025]: E1004 10:36:24.411473 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:24 crc kubenswrapper[5025]: I1004 10:36:24.411622 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:24 crc kubenswrapper[5025]: E1004 10:36:24.411678 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:25 crc kubenswrapper[5025]: I1004 10:36:25.410846 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:25 crc kubenswrapper[5025]: E1004 10:36:25.410991 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:26 crc kubenswrapper[5025]: I1004 10:36:26.410967 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:26 crc kubenswrapper[5025]: I1004 10:36:26.411074 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:26 crc kubenswrapper[5025]: I1004 10:36:26.411192 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:26 crc kubenswrapper[5025]: E1004 10:36:26.411199 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:26 crc kubenswrapper[5025]: E1004 10:36:26.411447 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:26 crc kubenswrapper[5025]: E1004 10:36:26.411812 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:27 crc kubenswrapper[5025]: I1004 10:36:27.130836 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n49fv_8d214492-7532-4f72-b032-b3cc1ae19473/kube-multus/1.log" Oct 04 10:36:27 crc kubenswrapper[5025]: I1004 10:36:27.131754 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n49fv_8d214492-7532-4f72-b032-b3cc1ae19473/kube-multus/0.log" Oct 04 10:36:27 crc kubenswrapper[5025]: I1004 10:36:27.131836 5025 generic.go:334] "Generic (PLEG): container finished" podID="8d214492-7532-4f72-b032-b3cc1ae19473" containerID="30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850" exitCode=1 Oct 04 10:36:27 crc kubenswrapper[5025]: I1004 10:36:27.131879 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n49fv" event={"ID":"8d214492-7532-4f72-b032-b3cc1ae19473","Type":"ContainerDied","Data":"30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850"} Oct 04 10:36:27 crc kubenswrapper[5025]: I1004 10:36:27.131933 5025 scope.go:117] "RemoveContainer" containerID="507080e97fd118d29790dbff51b92d4346f52aa94d46668966be48ff586cc172" Oct 04 10:36:27 crc kubenswrapper[5025]: I1004 10:36:27.132542 5025 scope.go:117] "RemoveContainer" containerID="30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850" Oct 04 10:36:27 crc kubenswrapper[5025]: E1004 10:36:27.133158 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-n49fv_openshift-multus(8d214492-7532-4f72-b032-b3cc1ae19473)\"" pod="openshift-multus/multus-n49fv" podUID="8d214492-7532-4f72-b032-b3cc1ae19473" Oct 04 10:36:27 crc kubenswrapper[5025]: I1004 10:36:27.410358 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:27 crc kubenswrapper[5025]: E1004 10:36:27.410535 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:28 crc kubenswrapper[5025]: I1004 10:36:28.138418 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n49fv_8d214492-7532-4f72-b032-b3cc1ae19473/kube-multus/1.log" Oct 04 10:36:28 crc kubenswrapper[5025]: I1004 10:36:28.410960 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:28 crc kubenswrapper[5025]: I1004 10:36:28.411099 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:28 crc kubenswrapper[5025]: E1004 10:36:28.411203 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:28 crc kubenswrapper[5025]: I1004 10:36:28.411116 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:28 crc kubenswrapper[5025]: E1004 10:36:28.411292 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:28 crc kubenswrapper[5025]: E1004 10:36:28.411427 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:29 crc kubenswrapper[5025]: I1004 10:36:29.411239 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:29 crc kubenswrapper[5025]: E1004 10:36:29.411425 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:30 crc kubenswrapper[5025]: I1004 10:36:30.410646 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:30 crc kubenswrapper[5025]: I1004 10:36:30.410724 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:30 crc kubenswrapper[5025]: E1004 10:36:30.410859 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:30 crc kubenswrapper[5025]: I1004 10:36:30.410880 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:30 crc kubenswrapper[5025]: E1004 10:36:30.411161 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:30 crc kubenswrapper[5025]: E1004 10:36:30.411363 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:31 crc kubenswrapper[5025]: I1004 10:36:31.410655 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:31 crc kubenswrapper[5025]: E1004 10:36:31.410840 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:32 crc kubenswrapper[5025]: E1004 10:36:32.389078 5025 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 04 10:36:32 crc kubenswrapper[5025]: I1004 10:36:32.411304 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:32 crc kubenswrapper[5025]: I1004 10:36:32.411314 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:32 crc kubenswrapper[5025]: E1004 10:36:32.411529 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:32 crc kubenswrapper[5025]: I1004 10:36:32.411553 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:32 crc kubenswrapper[5025]: E1004 10:36:32.411665 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:32 crc kubenswrapper[5025]: E1004 10:36:32.411898 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:32 crc kubenswrapper[5025]: E1004 10:36:32.548856 5025 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 10:36:33 crc kubenswrapper[5025]: I1004 10:36:33.410620 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:33 crc kubenswrapper[5025]: E1004 10:36:33.410751 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:34 crc kubenswrapper[5025]: I1004 10:36:34.410957 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:34 crc kubenswrapper[5025]: I1004 10:36:34.411177 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:34 crc kubenswrapper[5025]: E1004 10:36:34.411321 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:34 crc kubenswrapper[5025]: I1004 10:36:34.411367 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:34 crc kubenswrapper[5025]: E1004 10:36:34.411566 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:34 crc kubenswrapper[5025]: E1004 10:36:34.412168 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:34 crc kubenswrapper[5025]: I1004 10:36:34.412909 5025 scope.go:117] "RemoveContainer" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:36:35 crc kubenswrapper[5025]: I1004 10:36:35.165525 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/3.log" Oct 04 10:36:35 crc kubenswrapper[5025]: I1004 10:36:35.169234 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerStarted","Data":"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a"} Oct 04 10:36:35 crc kubenswrapper[5025]: I1004 10:36:35.170239 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:36:35 crc kubenswrapper[5025]: I1004 10:36:35.215141 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podStartSLOduration=102.215109384 podStartE2EDuration="1m42.215109384s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:35.213616174 +0000 UTC m=+123.638583084" watchObservedRunningTime="2025-10-04 10:36:35.215109384 +0000 UTC m=+123.640076314" Oct 04 10:36:35 crc kubenswrapper[5025]: I1004 10:36:35.293528 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-frc27"] Oct 04 10:36:35 crc kubenswrapper[5025]: I1004 10:36:35.293732 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:35 crc kubenswrapper[5025]: E1004 10:36:35.293926 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:36 crc kubenswrapper[5025]: I1004 10:36:36.410833 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:36 crc kubenswrapper[5025]: E1004 10:36:36.411382 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:36 crc kubenswrapper[5025]: I1004 10:36:36.410917 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:36 crc kubenswrapper[5025]: I1004 10:36:36.410917 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:36 crc kubenswrapper[5025]: E1004 10:36:36.411510 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:36 crc kubenswrapper[5025]: I1004 10:36:36.411555 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:36 crc kubenswrapper[5025]: E1004 10:36:36.411691 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:36 crc kubenswrapper[5025]: E1004 10:36:36.411772 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:37 crc kubenswrapper[5025]: E1004 10:36:37.549973 5025 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 04 10:36:38 crc kubenswrapper[5025]: I1004 10:36:38.410677 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:38 crc kubenswrapper[5025]: I1004 10:36:38.410725 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:38 crc kubenswrapper[5025]: E1004 10:36:38.410861 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:38 crc kubenswrapper[5025]: I1004 10:36:38.410686 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:38 crc kubenswrapper[5025]: I1004 10:36:38.410897 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:38 crc kubenswrapper[5025]: E1004 10:36:38.411153 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:38 crc kubenswrapper[5025]: E1004 10:36:38.411279 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:38 crc kubenswrapper[5025]: E1004 10:36:38.411554 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:38 crc kubenswrapper[5025]: I1004 10:36:38.411801 5025 scope.go:117] "RemoveContainer" containerID="30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850" Oct 04 10:36:39 crc kubenswrapper[5025]: I1004 10:36:39.188294 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n49fv_8d214492-7532-4f72-b032-b3cc1ae19473/kube-multus/1.log" Oct 04 10:36:39 crc kubenswrapper[5025]: I1004 10:36:39.188693 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n49fv" event={"ID":"8d214492-7532-4f72-b032-b3cc1ae19473","Type":"ContainerStarted","Data":"f1d8ee42e06786ad69717a96ff5cdc23094d6c2bb0f8a56b8a35ed098ba52011"} Oct 04 10:36:40 crc kubenswrapper[5025]: I1004 10:36:40.410561 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:40 crc kubenswrapper[5025]: E1004 10:36:40.410743 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:40 crc kubenswrapper[5025]: I1004 10:36:40.411107 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:40 crc kubenswrapper[5025]: I1004 10:36:40.411303 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:40 crc kubenswrapper[5025]: E1004 10:36:40.411365 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:40 crc kubenswrapper[5025]: I1004 10:36:40.411624 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:40 crc kubenswrapper[5025]: E1004 10:36:40.411919 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:40 crc kubenswrapper[5025]: E1004 10:36:40.412347 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:42 crc kubenswrapper[5025]: I1004 10:36:42.410215 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:42 crc kubenswrapper[5025]: I1004 10:36:42.410250 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:42 crc kubenswrapper[5025]: I1004 10:36:42.410265 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:42 crc kubenswrapper[5025]: I1004 10:36:42.410305 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:42 crc kubenswrapper[5025]: E1004 10:36:42.412380 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-frc27" podUID="86312d38-63d0-409a-98d4-727f0fb47929" Oct 04 10:36:42 crc kubenswrapper[5025]: E1004 10:36:42.412501 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 04 10:36:42 crc kubenswrapper[5025]: E1004 10:36:42.412584 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 04 10:36:42 crc kubenswrapper[5025]: E1004 10:36:42.412641 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 04 10:36:44 crc kubenswrapper[5025]: I1004 10:36:44.410390 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:36:44 crc kubenswrapper[5025]: I1004 10:36:44.410468 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:36:44 crc kubenswrapper[5025]: I1004 10:36:44.410518 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:36:44 crc kubenswrapper[5025]: I1004 10:36:44.412094 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:36:44 crc kubenswrapper[5025]: I1004 10:36:44.413567 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 04 10:36:44 crc kubenswrapper[5025]: I1004 10:36:44.415228 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 04 10:36:44 crc kubenswrapper[5025]: I1004 10:36:44.415493 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 04 10:36:44 crc kubenswrapper[5025]: I1004 10:36:44.415745 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 04 10:36:44 crc kubenswrapper[5025]: I1004 10:36:44.416670 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 04 10:36:44 crc kubenswrapper[5025]: I1004 10:36:44.417938 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.908770 5025 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.961100 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2j4wq"] Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.962117 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.962403 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk"] Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.963831 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.969594 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn"] Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.970379 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.971941 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk"] Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.972323 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.973487 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.974791 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.974790 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6mdn4"] Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.975509 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.975709 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.976274 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.977716 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.978087 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.978408 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg"] Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.978535 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.978603 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.978749 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.979092 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.979186 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.981284 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rxmh4"] Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.982407 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.986214 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.992554 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.992756 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 10:36:50 crc kubenswrapper[5025]: I1004 10:36:50.993273 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.001550 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.003843 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.004189 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.004347 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.004805 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.004895 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.005691 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.005707 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.006103 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.006579 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.007218 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.007388 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.007421 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.007613 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.008493 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.008642 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.008884 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.009050 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.009590 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.010148 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.010357 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.010516 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.015069 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.015296 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.015449 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.015684 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.015893 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.015984 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.016286 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.016404 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.016497 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.016622 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.016775 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.016978 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.017165 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.019787 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mlpcc"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.020121 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.020422 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-mvbs7"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.025803 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.029269 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z7zzf"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.030773 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.031366 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mvbs7" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.049850 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.050219 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.050236 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.051075 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5bv89"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.051124 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.051964 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.052772 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jbffs"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.054037 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.057616 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.057813 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.057973 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.058249 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.058878 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.058901 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.059260 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.059440 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.059595 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.059923 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.059955 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.060191 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.060322 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.060382 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.060874 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.061010 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.061141 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.061498 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.061649 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.062073 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.074848 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/227aafde-cb49-4cff-b561-d7ba6b84a1cc-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.074912 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ada2393-a845-408e-b527-e26c7af5b037-serving-cert\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.074937 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7sd4\" (UniqueName: \"kubernetes.io/projected/78804be3-9972-43a8-9087-1e7b62c96f91-kube-api-access-g7sd4\") pod \"cluster-samples-operator-665b6dd947-t8lsk\" (UID: \"78804be3-9972-43a8-9087-1e7b62c96f91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.074958 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-client-ca\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.074976 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/eeae56c3-931d-4b7d-8941-53152690a562-machine-approver-tls\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075000 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeae56c3-931d-4b7d-8941-53152690a562-config\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075056 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b33f1ca5-6029-4855-96eb-f13d7fd333f8-etcd-client\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075075 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b33f1ca5-6029-4855-96eb-f13d7fd333f8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075100 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33f1ca5-6029-4855-96eb-f13d7fd333f8-serving-cert\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075122 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b33f1ca5-6029-4855-96eb-f13d7fd333f8-encryption-config\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075149 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-client-ca\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075170 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b33f1ca5-6029-4855-96eb-f13d7fd333f8-audit-policies\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075190 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whbq2\" (UniqueName: \"kubernetes.io/projected/b33f1ca5-6029-4855-96eb-f13d7fd333f8-kube-api-access-whbq2\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075218 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8wvl\" (UniqueName: \"kubernetes.io/projected/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-kube-api-access-q8wvl\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075247 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7nh9\" (UniqueName: \"kubernetes.io/projected/eeae56c3-931d-4b7d-8941-53152690a562-kube-api-access-p7nh9\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075266 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/227aafde-cb49-4cff-b561-d7ba6b84a1cc-config\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075290 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-serving-cert\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075311 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/227aafde-cb49-4cff-b561-d7ba6b84a1cc-images\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075345 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fczqh\" (UniqueName: \"kubernetes.io/projected/227aafde-cb49-4cff-b561-d7ba6b84a1cc-kube-api-access-fczqh\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075373 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68hm6\" (UniqueName: \"kubernetes.io/projected/9ada2393-a845-408e-b527-e26c7af5b037-kube-api-access-68hm6\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075393 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/78804be3-9972-43a8-9087-1e7b62c96f91-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-t8lsk\" (UID: \"78804be3-9972-43a8-9087-1e7b62c96f91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075414 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-config\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075437 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075457 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-config\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075482 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b33f1ca5-6029-4855-96eb-f13d7fd333f8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075508 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b33f1ca5-6029-4855-96eb-f13d7fd333f8-audit-dir\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.075532 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eeae56c3-931d-4b7d-8941-53152690a562-auth-proxy-config\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.078636 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.079657 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.079706 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.080396 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.081280 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.081532 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.081841 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.082116 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.082285 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.082565 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.082740 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.083965 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.086093 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzljn"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.086671 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.091840 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.091857 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.095415 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.095884 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.095978 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.106824 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.106993 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.107212 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.107380 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.106831 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.108524 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvzjs"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.109130 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.115346 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.115779 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.120463 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.122439 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5bv89"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.127169 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.127440 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.127562 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.132225 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.132418 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.132539 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.132623 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.132692 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.132906 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.133197 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.133279 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.134305 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6mdn4"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.136884 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.141928 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.149653 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.150487 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.150884 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.151647 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.151795 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176412 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-trusted-ca-bundle\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176478 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-client-ca\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176514 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whbq2\" (UniqueName: \"kubernetes.io/projected/b33f1ca5-6029-4855-96eb-f13d7fd333f8-kube-api-access-whbq2\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176546 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b33f1ca5-6029-4855-96eb-f13d7fd333f8-audit-policies\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176575 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1613b097-9c6f-4ae1-84b3-fc12833ee019-metrics-tls\") pod \"dns-operator-744455d44c-z7zzf\" (UID: \"1613b097-9c6f-4ae1-84b3-fc12833ee019\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176607 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-oauth-serving-cert\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176644 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/003ddd1f-9614-417f-b8d9-e6efdc75b922-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9bbj8\" (UID: \"003ddd1f-9614-417f-b8d9-e6efdc75b922\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176680 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8wvl\" (UniqueName: \"kubernetes.io/projected/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-kube-api-access-q8wvl\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176710 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-audit\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176739 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d92c5448-95d8-4008-a5a0-2c1a5a016db3-etcd-client\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176771 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7nh9\" (UniqueName: \"kubernetes.io/projected/eeae56c3-931d-4b7d-8941-53152690a562-kube-api-access-p7nh9\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176802 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/227aafde-cb49-4cff-b561-d7ba6b84a1cc-config\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176833 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d92c5448-95d8-4008-a5a0-2c1a5a016db3-etcd-service-ca\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176868 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-serving-cert\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176896 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/227aafde-cb49-4cff-b561-d7ba6b84a1cc-images\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176922 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrqmf\" (UniqueName: \"kubernetes.io/projected/1613b097-9c6f-4ae1-84b3-fc12833ee019-kube-api-access-rrqmf\") pod \"dns-operator-744455d44c-z7zzf\" (UID: \"1613b097-9c6f-4ae1-84b3-fc12833ee019\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.176959 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-trusted-ca\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177031 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thth5\" (UniqueName: \"kubernetes.io/projected/dacae947-fe70-4eef-8f0c-8f012cea7c96-kube-api-access-thth5\") pod \"openshift-apiserver-operator-796bbdcf4f-8kd6s\" (UID: \"dacae947-fe70-4eef-8f0c-8f012cea7c96\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177062 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ed7f7173-bdd4-4ef2-a780-48c87a23125b-node-pullsecrets\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177131 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fczqh\" (UniqueName: \"kubernetes.io/projected/227aafde-cb49-4cff-b561-d7ba6b84a1cc-kube-api-access-fczqh\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177164 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-service-ca\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177195 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-serving-cert\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177224 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/78804be3-9972-43a8-9087-1e7b62c96f91-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-t8lsk\" (UID: \"78804be3-9972-43a8-9087-1e7b62c96f91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177253 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-config\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177281 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb222f50-1834-4c9f-a6da-4d1bbd383398-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177313 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68hm6\" (UniqueName: \"kubernetes.io/projected/9ada2393-a845-408e-b527-e26c7af5b037-kube-api-access-68hm6\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177337 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177366 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-config\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177396 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177434 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b33f1ca5-6029-4855-96eb-f13d7fd333f8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177466 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d92c5448-95d8-4008-a5a0-2c1a5a016db3-config\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177500 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrppw\" (UniqueName: \"kubernetes.io/projected/003ddd1f-9614-417f-b8d9-e6efdc75b922-kube-api-access-xrppw\") pod \"openshift-config-operator-7777fb866f-9bbj8\" (UID: \"003ddd1f-9614-417f-b8d9-e6efdc75b922\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177532 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-config\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177566 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b33f1ca5-6029-4855-96eb-f13d7fd333f8-audit-dir\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177599 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eeae56c3-931d-4b7d-8941-53152690a562-auth-proxy-config\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177629 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fb222f50-1834-4c9f-a6da-4d1bbd383398-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177663 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d92c5448-95d8-4008-a5a0-2c1a5a016db3-etcd-ca\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177724 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-image-import-ca\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177794 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacae947-fe70-4eef-8f0c-8f012cea7c96-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8kd6s\" (UID: \"dacae947-fe70-4eef-8f0c-8f012cea7c96\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177883 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ed7f7173-bdd4-4ef2-a780-48c87a23125b-audit-dir\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177937 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/227aafde-cb49-4cff-b561-d7ba6b84a1cc-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.177972 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-etcd-serving-ca\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178038 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-console-config\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178073 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ed7f7173-bdd4-4ef2-a780-48c87a23125b-encryption-config\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178105 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ada2393-a845-408e-b527-e26c7af5b037-serving-cert\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178138 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7sd4\" (UniqueName: \"kubernetes.io/projected/78804be3-9972-43a8-9087-1e7b62c96f91-kube-api-access-g7sd4\") pod \"cluster-samples-operator-665b6dd947-t8lsk\" (UID: \"78804be3-9972-43a8-9087-1e7b62c96f91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178171 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr97x\" (UniqueName: \"kubernetes.io/projected/0f85106f-c448-496a-8199-9a03990115ac-kube-api-access-pr97x\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178233 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacae947-fe70-4eef-8f0c-8f012cea7c96-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8kd6s\" (UID: \"dacae947-fe70-4eef-8f0c-8f012cea7c96\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178267 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb222f50-1834-4c9f-a6da-4d1bbd383398-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178302 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trcds\" (UniqueName: \"kubernetes.io/projected/ed7f7173-bdd4-4ef2-a780-48c87a23125b-kube-api-access-trcds\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178339 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-client-ca\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178375 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/eeae56c3-931d-4b7d-8941-53152690a562-machine-approver-tls\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178411 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeae56c3-931d-4b7d-8941-53152690a562-config\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178449 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2mq4\" (UniqueName: \"kubernetes.io/projected/fb222f50-1834-4c9f-a6da-4d1bbd383398-kube-api-access-q2mq4\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178486 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b33f1ca5-6029-4855-96eb-f13d7fd333f8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178554 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/003ddd1f-9614-417f-b8d9-e6efdc75b922-serving-cert\") pod \"openshift-config-operator-7777fb866f-9bbj8\" (UID: \"003ddd1f-9614-417f-b8d9-e6efdc75b922\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178599 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed7f7173-bdd4-4ef2-a780-48c87a23125b-etcd-client\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178657 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b33f1ca5-6029-4855-96eb-f13d7fd333f8-etcd-client\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178690 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-oauth-config\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178726 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c6bh\" (UniqueName: \"kubernetes.io/projected/f42407c5-6c5a-452c-8ad9-6ab593111376-kube-api-access-7c6bh\") pod \"downloads-7954f5f757-mvbs7\" (UID: \"f42407c5-6c5a-452c-8ad9-6ab593111376\") " pod="openshift-console/downloads-7954f5f757-mvbs7" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178783 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33f1ca5-6029-4855-96eb-f13d7fd333f8-serving-cert\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178815 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed7f7173-bdd4-4ef2-a780-48c87a23125b-serving-cert\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178842 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-config\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.178873 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d92c5448-95d8-4008-a5a0-2c1a5a016db3-serving-cert\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.180184 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-serving-cert\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.180239 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dtsg\" (UniqueName: \"kubernetes.io/projected/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-kube-api-access-5dtsg\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.180266 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf8lz\" (UniqueName: \"kubernetes.io/projected/d92c5448-95d8-4008-a5a0-2c1a5a016db3-kube-api-access-cf8lz\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.180302 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b33f1ca5-6029-4855-96eb-f13d7fd333f8-encryption-config\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.185093 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mlpcc"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.185157 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rxmh4"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.186558 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.187934 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-client-ca\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.189088 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2j4wq"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.188861 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/eeae56c3-931d-4b7d-8941-53152690a562-auth-proxy-config\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.189309 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/227aafde-cb49-4cff-b561-d7ba6b84a1cc-images\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.190200 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-client-ca\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.190335 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b33f1ca5-6029-4855-96eb-f13d7fd333f8-audit-dir\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.190566 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.190664 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eeae56c3-931d-4b7d-8941-53152690a562-config\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.191533 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b33f1ca5-6029-4855-96eb-f13d7fd333f8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.191729 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-config\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.192358 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/227aafde-cb49-4cff-b561-d7ba6b84a1cc-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.192373 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ada2393-a845-408e-b527-e26c7af5b037-serving-cert\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.193164 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.193320 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b33f1ca5-6029-4855-96eb-f13d7fd333f8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.193481 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/b33f1ca5-6029-4855-96eb-f13d7fd333f8-audit-policies\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.194158 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/227aafde-cb49-4cff-b561-d7ba6b84a1cc-config\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.195030 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.195263 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.195853 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b33f1ca5-6029-4855-96eb-f13d7fd333f8-etcd-client\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.196356 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-serving-cert\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.196749 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b33f1ca5-6029-4855-96eb-f13d7fd333f8-encryption-config\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.197379 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/78804be3-9972-43a8-9087-1e7b62c96f91-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-t8lsk\" (UID: \"78804be3-9972-43a8-9087-1e7b62c96f91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.200091 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.200470 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-config\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.200883 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/eeae56c3-931d-4b7d-8941-53152690a562-machine-approver-tls\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.202133 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b33f1ca5-6029-4855-96eb-f13d7fd333f8-serving-cert\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.205634 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.205942 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2m4s6"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.206883 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.219992 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.220570 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-gq2bg"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.220930 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gpd22"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.221892 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.222334 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.224224 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.238297 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.238409 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.238580 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.238720 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.238881 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.238980 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.239167 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.239188 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.239306 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.239551 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.239895 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.240333 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.240499 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.240626 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.241043 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.241502 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.244234 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.248400 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.248941 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.249556 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hssvl"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.249933 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.250157 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.250277 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.254893 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.255409 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.255873 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fnszf"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.257691 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sp4mz"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.257850 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.258056 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.258197 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.258489 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.259957 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.260483 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.260819 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.260996 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.264580 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7mw6v"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.265613 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.265657 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzljn"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.267121 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.268607 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2m4s6"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.269526 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gpd22"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.270579 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.272755 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.273766 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.274716 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.277087 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mvbs7"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.278034 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.279196 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-bn5rg"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.279750 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bn5rg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.281156 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hssvl"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.281456 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d92c5448-95d8-4008-a5a0-2c1a5a016db3-etcd-ca\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.281527 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fb222f50-1834-4c9f-a6da-4d1bbd383398-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.281634 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-image-import-ca\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.281710 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacae947-fe70-4eef-8f0c-8f012cea7c96-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8kd6s\" (UID: \"dacae947-fe70-4eef-8f0c-8f012cea7c96\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.281792 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-etcd-serving-ca\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.281838 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ed7f7173-bdd4-4ef2-a780-48c87a23125b-audit-dir\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.281916 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-console-config\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.281975 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr97x\" (UniqueName: \"kubernetes.io/projected/0f85106f-c448-496a-8199-9a03990115ac-kube-api-access-pr97x\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282008 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacae947-fe70-4eef-8f0c-8f012cea7c96-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8kd6s\" (UID: \"dacae947-fe70-4eef-8f0c-8f012cea7c96\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282074 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ed7f7173-bdd4-4ef2-a780-48c87a23125b-encryption-config\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282125 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb222f50-1834-4c9f-a6da-4d1bbd383398-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282173 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trcds\" (UniqueName: \"kubernetes.io/projected/ed7f7173-bdd4-4ef2-a780-48c87a23125b-kube-api-access-trcds\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282229 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2mq4\" (UniqueName: \"kubernetes.io/projected/fb222f50-1834-4c9f-a6da-4d1bbd383398-kube-api-access-q2mq4\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282294 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/003ddd1f-9614-417f-b8d9-e6efdc75b922-serving-cert\") pod \"openshift-config-operator-7777fb866f-9bbj8\" (UID: \"003ddd1f-9614-417f-b8d9-e6efdc75b922\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282340 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed7f7173-bdd4-4ef2-a780-48c87a23125b-etcd-client\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282370 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-oauth-config\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282411 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed7f7173-bdd4-4ef2-a780-48c87a23125b-serving-cert\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282450 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c6bh\" (UniqueName: \"kubernetes.io/projected/f42407c5-6c5a-452c-8ad9-6ab593111376-kube-api-access-7c6bh\") pod \"downloads-7954f5f757-mvbs7\" (UID: \"f42407c5-6c5a-452c-8ad9-6ab593111376\") " pod="openshift-console/downloads-7954f5f757-mvbs7" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282490 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-config\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282520 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d92c5448-95d8-4008-a5a0-2c1a5a016db3-serving-cert\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282561 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-serving-cert\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282596 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dtsg\" (UniqueName: \"kubernetes.io/projected/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-kube-api-access-5dtsg\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282631 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf8lz\" (UniqueName: \"kubernetes.io/projected/d92c5448-95d8-4008-a5a0-2c1a5a016db3-kube-api-access-cf8lz\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282668 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-trusted-ca-bundle\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282741 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1613b097-9c6f-4ae1-84b3-fc12833ee019-metrics-tls\") pod \"dns-operator-744455d44c-z7zzf\" (UID: \"1613b097-9c6f-4ae1-84b3-fc12833ee019\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282785 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-oauth-serving-cert\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282829 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/003ddd1f-9614-417f-b8d9-e6efdc75b922-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9bbj8\" (UID: \"003ddd1f-9614-417f-b8d9-e6efdc75b922\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282877 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-audit\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282907 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d92c5448-95d8-4008-a5a0-2c1a5a016db3-etcd-client\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282967 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dacae947-fe70-4eef-8f0c-8f012cea7c96-config\") pod \"openshift-apiserver-operator-796bbdcf4f-8kd6s\" (UID: \"dacae947-fe70-4eef-8f0c-8f012cea7c96\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.283061 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.282960 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d92c5448-95d8-4008-a5a0-2c1a5a016db3-etcd-service-ca\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.283165 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrqmf\" (UniqueName: \"kubernetes.io/projected/1613b097-9c6f-4ae1-84b3-fc12833ee019-kube-api-access-rrqmf\") pod \"dns-operator-744455d44c-z7zzf\" (UID: \"1613b097-9c6f-4ae1-84b3-fc12833ee019\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.283206 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-trusted-ca\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.283240 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thth5\" (UniqueName: \"kubernetes.io/projected/dacae947-fe70-4eef-8f0c-8f012cea7c96-kube-api-access-thth5\") pod \"openshift-apiserver-operator-796bbdcf4f-8kd6s\" (UID: \"dacae947-fe70-4eef-8f0c-8f012cea7c96\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.283279 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ed7f7173-bdd4-4ef2-a780-48c87a23125b-node-pullsecrets\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.283337 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-image-import-ca\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.283391 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-serving-cert\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.283425 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-service-ca\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.283466 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb222f50-1834-4c9f-a6da-4d1bbd383398-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.283501 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.284762 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrppw\" (UniqueName: \"kubernetes.io/projected/003ddd1f-9614-417f-b8d9-e6efdc75b922-kube-api-access-xrppw\") pod \"openshift-config-operator-7777fb866f-9bbj8\" (UID: \"003ddd1f-9614-417f-b8d9-e6efdc75b922\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.284816 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-config\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.284881 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d92c5448-95d8-4008-a5a0-2c1a5a016db3-config\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.285153 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/003ddd1f-9614-417f-b8d9-e6efdc75b922-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9bbj8\" (UID: \"003ddd1f-9614-417f-b8d9-e6efdc75b922\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.285487 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/d92c5448-95d8-4008-a5a0-2c1a5a016db3-etcd-ca\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.285523 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-etcd-serving-ca\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.285773 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb222f50-1834-4c9f-a6da-4d1bbd383398-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.285945 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d92c5448-95d8-4008-a5a0-2c1a5a016db3-config\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.286637 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dacae947-fe70-4eef-8f0c-8f012cea7c96-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-8kd6s\" (UID: \"dacae947-fe70-4eef-8f0c-8f012cea7c96\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.288549 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-config\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.288060 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ed7f7173-bdd4-4ef2-a780-48c87a23125b-audit-dir\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.289669 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-console-config\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.289705 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d92c5448-95d8-4008-a5a0-2c1a5a016db3-serving-cert\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.289948 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d92c5448-95d8-4008-a5a0-2c1a5a016db3-etcd-client\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.291071 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-trusted-ca-bundle\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.291852 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-audit\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.291933 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.292196 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-oauth-config\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.292541 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-config\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.293076 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-oauth-serving-cert\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.293843 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fb222f50-1834-4c9f-a6da-4d1bbd383398-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.293899 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/ed7f7173-bdd4-4ef2-a780-48c87a23125b-node-pullsecrets\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.294314 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-service-ca\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.294623 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-trusted-ca\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.295232 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ed7f7173-bdd4-4ef2-a780-48c87a23125b-etcd-client\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.295330 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1613b097-9c6f-4ae1-84b3-fc12833ee019-metrics-tls\") pod \"dns-operator-744455d44c-z7zzf\" (UID: \"1613b097-9c6f-4ae1-84b3-fc12833ee019\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.295705 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/d92c5448-95d8-4008-a5a0-2c1a5a016db3-etcd-service-ca\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.295814 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed7f7173-bdd4-4ef2-a780-48c87a23125b-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.295911 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/003ddd1f-9614-417f-b8d9-e6efdc75b922-serving-cert\") pod \"openshift-config-operator-7777fb866f-9bbj8\" (UID: \"003ddd1f-9614-417f-b8d9-e6efdc75b922\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.296157 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7mw6v"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.299140 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jbffs"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.299182 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ed7f7173-bdd4-4ef2-a780-48c87a23125b-encryption-config\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.299456 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ed7f7173-bdd4-4ef2-a780-48c87a23125b-serving-cert\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.300769 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-serving-cert\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.302530 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-serving-cert\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.303095 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.304316 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z7zzf"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.307284 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.310083 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvzjs"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.311354 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.313261 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.314759 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-fzxjq"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.315730 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.316305 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-zjclt"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.317330 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.317682 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whbq2\" (UniqueName: \"kubernetes.io/projected/b33f1ca5-6029-4855-96eb-f13d7fd333f8-kube-api-access-whbq2\") pod \"apiserver-7bbb656c7d-qlhdg\" (UID: \"b33f1ca5-6029-4855-96eb-f13d7fd333f8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.317905 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.319063 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bn5rg"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.320097 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fnszf"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.321201 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sp4mz"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.322375 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.323374 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.324464 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zjclt"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.336536 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7sd4\" (UniqueName: \"kubernetes.io/projected/78804be3-9972-43a8-9087-1e7b62c96f91-kube-api-access-g7sd4\") pod \"cluster-samples-operator-665b6dd947-t8lsk\" (UID: \"78804be3-9972-43a8-9087-1e7b62c96f91\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.356391 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fczqh\" (UniqueName: \"kubernetes.io/projected/227aafde-cb49-4cff-b561-d7ba6b84a1cc-kube-api-access-fczqh\") pod \"machine-api-operator-5694c8668f-6mdn4\" (UID: \"227aafde-cb49-4cff-b561-d7ba6b84a1cc\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.360340 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.379527 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68hm6\" (UniqueName: \"kubernetes.io/projected/9ada2393-a845-408e-b527-e26c7af5b037-kube-api-access-68hm6\") pod \"route-controller-manager-6576b87f9c-rbctn\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.393202 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.405342 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8wvl\" (UniqueName: \"kubernetes.io/projected/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-kube-api-access-q8wvl\") pod \"controller-manager-879f6c89f-2j4wq\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.421510 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7nh9\" (UniqueName: \"kubernetes.io/projected/eeae56c3-931d-4b7d-8941-53152690a562-kube-api-access-p7nh9\") pod \"machine-approver-56656f9798-bpqmk\" (UID: \"eeae56c3-931d-4b7d-8941-53152690a562\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.424079 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.451881 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.469505 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.491344 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.510403 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.523978 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.545493 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.565375 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.573655 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6mdn4"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.585974 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.599648 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.612342 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.613898 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.614943 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.619987 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.624447 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 04 10:36:51 crc kubenswrapper[5025]: W1004 10:36:51.628062 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb33f1ca5_6029_4855_96eb_f13d7fd333f8.slice/crio-aa6ee600f066cc7644d6db37346e47c1cd090c95e507f9bd9eadbc36d1b03014 WatchSource:0}: Error finding container aa6ee600f066cc7644d6db37346e47c1cd090c95e507f9bd9eadbc36d1b03014: Status 404 returned error can't find the container with id aa6ee600f066cc7644d6db37346e47c1cd090c95e507f9bd9eadbc36d1b03014 Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.647483 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.650028 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.664118 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.685526 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.705583 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.726977 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.747660 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.766209 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.785479 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.807340 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.813055 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2j4wq"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.824615 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.829287 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.844739 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: W1004 10:36:51.852554 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ada2393_a845_408e_b527_e26c7af5b037.slice/crio-29de14d693ac83733e88730a5fee52883e5522fcb32389e929f99684c31f333b WatchSource:0}: Error finding container 29de14d693ac83733e88730a5fee52883e5522fcb32389e929f99684c31f333b: Status 404 returned error can't find the container with id 29de14d693ac83733e88730a5fee52883e5522fcb32389e929f99684c31f333b Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.865300 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.867007 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk"] Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.884998 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.904189 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.925588 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.945001 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.964916 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 04 10:36:51 crc kubenswrapper[5025]: I1004 10:36:51.984617 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.004311 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.023920 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.044838 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.063918 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.083866 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.104439 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.124559 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.144183 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.163516 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.184063 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.206113 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.224963 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.238398 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" event={"ID":"9ada2393-a845-408e-b527-e26c7af5b037","Type":"ContainerStarted","Data":"91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.238471 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" event={"ID":"9ada2393-a845-408e-b527-e26c7af5b037","Type":"ContainerStarted","Data":"29de14d693ac83733e88730a5fee52883e5522fcb32389e929f99684c31f333b"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.239700 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.241995 5025 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-rbctn container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.242085 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" podUID="9ada2393-a845-408e-b527-e26c7af5b037" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.242560 5025 request.go:700] Waited for 1.001736075s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.242641 5025 generic.go:334] "Generic (PLEG): container finished" podID="b33f1ca5-6029-4855-96eb-f13d7fd333f8" containerID="20a728662bf78119363406a241053cc8bc517ceb4df4f7df229de2c1f7458509" exitCode=0 Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.242791 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" event={"ID":"b33f1ca5-6029-4855-96eb-f13d7fd333f8","Type":"ContainerDied","Data":"20a728662bf78119363406a241053cc8bc517ceb4df4f7df229de2c1f7458509"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.242831 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" event={"ID":"b33f1ca5-6029-4855-96eb-f13d7fd333f8","Type":"ContainerStarted","Data":"aa6ee600f066cc7644d6db37346e47c1cd090c95e507f9bd9eadbc36d1b03014"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.244126 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.247397 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" event={"ID":"227aafde-cb49-4cff-b561-d7ba6b84a1cc","Type":"ContainerStarted","Data":"b31c23426b46301d88c9829f03280bf6e06d5faa18e321a9d1f354a7ac8ca67f"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.247420 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" event={"ID":"227aafde-cb49-4cff-b561-d7ba6b84a1cc","Type":"ContainerStarted","Data":"d855997ed74cc82c4d9b2d24fbe51d0cf002f9a05d5dd3c08c9eab19276dedf1"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.247429 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" event={"ID":"227aafde-cb49-4cff-b561-d7ba6b84a1cc","Type":"ContainerStarted","Data":"738f4017e78f97b0b296f77056b0b798d40b0ec53c914dd47da6c0d5853843d8"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.249871 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" event={"ID":"78804be3-9972-43a8-9087-1e7b62c96f91","Type":"ContainerStarted","Data":"4c948fc5c55f8479f29a8e80f119b46346f1d7a7ff9cf9e09448f5f51229d799"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.249895 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" event={"ID":"78804be3-9972-43a8-9087-1e7b62c96f91","Type":"ContainerStarted","Data":"9648b78140a4409946882a7c99419f1452dc538c72eacc37f81800abd6e73f4c"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.251289 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" event={"ID":"df6dbb5c-fc0d-4020-af0e-f05bbd44949f","Type":"ContainerStarted","Data":"618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.251309 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" event={"ID":"df6dbb5c-fc0d-4020-af0e-f05bbd44949f","Type":"ContainerStarted","Data":"26ee863bfc65c7286e87ffac56a29aeec821f27276962d950a9231777e521d37"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.251606 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.254502 5025 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2j4wq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.254981 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" podUID="df6dbb5c-fc0d-4020-af0e-f05bbd44949f" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.257697 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" event={"ID":"eeae56c3-931d-4b7d-8941-53152690a562","Type":"ContainerStarted","Data":"097295745bc01186430028cafb259054cfddb4597bc6e5fc613b28073159c067"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.257743 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" event={"ID":"eeae56c3-931d-4b7d-8941-53152690a562","Type":"ContainerStarted","Data":"12bcc16423bc2d6ba80916ad400ea5355fa77001b817f06598200bfcbb39ab5e"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.257784 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" event={"ID":"eeae56c3-931d-4b7d-8941-53152690a562","Type":"ContainerStarted","Data":"7affa5aaf21cccdb1bd43180a3eb601d441deb120544b4cf4b66946dba62004f"} Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.264698 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.284634 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.304469 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.324724 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.344941 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.364650 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.384807 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.405697 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.426856 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.444988 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.464563 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.484698 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.504338 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.524143 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.544982 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.565904 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.584010 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.604287 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.624367 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.644977 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.664592 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.685790 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.711675 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.724302 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.744112 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.768509 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.784000 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.804668 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.823855 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.844218 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.863849 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.904258 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.924292 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.944965 5025 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.964669 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 04 10:36:52 crc kubenswrapper[5025]: I1004 10:36:52.985130 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.004542 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.025363 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.067519 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fb222f50-1834-4c9f-a6da-4d1bbd383398-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.078547 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2mq4\" (UniqueName: \"kubernetes.io/projected/fb222f50-1834-4c9f-a6da-4d1bbd383398-kube-api-access-q2mq4\") pod \"cluster-image-registry-operator-dc59b4c8b-zcqx6\" (UID: \"fb222f50-1834-4c9f-a6da-4d1bbd383398\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.103056 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trcds\" (UniqueName: \"kubernetes.io/projected/ed7f7173-bdd4-4ef2-a780-48c87a23125b-kube-api-access-trcds\") pod \"apiserver-76f77b778f-rxmh4\" (UID: \"ed7f7173-bdd4-4ef2-a780-48c87a23125b\") " pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.126347 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr97x\" (UniqueName: \"kubernetes.io/projected/0f85106f-c448-496a-8199-9a03990115ac-kube-api-access-pr97x\") pod \"console-f9d7485db-5bv89\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.140386 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c6bh\" (UniqueName: \"kubernetes.io/projected/f42407c5-6c5a-452c-8ad9-6ab593111376-kube-api-access-7c6bh\") pod \"downloads-7954f5f757-mvbs7\" (UID: \"f42407c5-6c5a-452c-8ad9-6ab593111376\") " pod="openshift-console/downloads-7954f5f757-mvbs7" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.163439 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrqmf\" (UniqueName: \"kubernetes.io/projected/1613b097-9c6f-4ae1-84b3-fc12833ee019-kube-api-access-rrqmf\") pod \"dns-operator-744455d44c-z7zzf\" (UID: \"1613b097-9c6f-4ae1-84b3-fc12833ee019\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.206523 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dtsg\" (UniqueName: \"kubernetes.io/projected/ba4c3d7f-46dc-4262-9c8a-0b2991ce7120-kube-api-access-5dtsg\") pod \"console-operator-58897d9998-jbffs\" (UID: \"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120\") " pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.213601 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.213902 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.214099 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.214743 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-policies\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.215560 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/acc502df-2c80-40fc-975f-d993333b1898-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.215601 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-registry-certificates\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.215871 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a726141a-3bd2-41d5-9c11-003b460b0716-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wrf9q\" (UID: \"a726141a-3bd2-41d5-9c11-003b460b0716\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.216002 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/acc502df-2c80-40fc-975f-d993333b1898-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.216058 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-dir\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.216139 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.216242 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.217419 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.217807 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-registry-tls\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.218087 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-bound-sa-token\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.218125 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.218159 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g7tl\" (UniqueName: \"kubernetes.io/projected/7a2c8998-2de1-453a-9adb-72facdc8bd2f-kube-api-access-9g7tl\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.218641 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltskc\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-kube-api-access-ltskc\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.219143 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-542nt\" (UniqueName: \"kubernetes.io/projected/a726141a-3bd2-41d5-9c11-003b460b0716-kube-api-access-542nt\") pod \"openshift-controller-manager-operator-756b6f6bc6-wrf9q\" (UID: \"a726141a-3bd2-41d5-9c11-003b460b0716\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.219200 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a726141a-3bd2-41d5-9c11-003b460b0716-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wrf9q\" (UID: \"a726141a-3bd2-41d5-9c11-003b460b0716\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.221053 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.221612 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.221978 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-trusted-ca\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.222076 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.222103 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.222131 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.222159 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.222483 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:53.72246308 +0000 UTC m=+142.147430150 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.234852 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrppw\" (UniqueName: \"kubernetes.io/projected/003ddd1f-9614-417f-b8d9-e6efdc75b922-kube-api-access-xrppw\") pod \"openshift-config-operator-7777fb866f-9bbj8\" (UID: \"003ddd1f-9614-417f-b8d9-e6efdc75b922\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.243002 5025 request.go:700] Waited for 1.948238994s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-etcd-operator/serviceaccounts/etcd-operator/token Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.248505 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.249273 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thth5\" (UniqueName: \"kubernetes.io/projected/dacae947-fe70-4eef-8f0c-8f012cea7c96-kube-api-access-thth5\") pod \"openshift-apiserver-operator-796bbdcf4f-8kd6s\" (UID: \"dacae947-fe70-4eef-8f0c-8f012cea7c96\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.264902 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.266252 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf8lz\" (UniqueName: \"kubernetes.io/projected/d92c5448-95d8-4008-a5a0-2c1a5a016db3-kube-api-access-cf8lz\") pod \"etcd-operator-b45778765-mlpcc\" (UID: \"d92c5448-95d8-4008-a5a0-2c1a5a016db3\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.269640 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" event={"ID":"78804be3-9972-43a8-9087-1e7b62c96f91","Type":"ContainerStarted","Data":"b27e4ad8683524c6b85123033f64f3d81e27bd685922b33dcc3f746b2825331b"} Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.277268 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" event={"ID":"b33f1ca5-6029-4855-96eb-f13d7fd333f8","Type":"ContainerStarted","Data":"ee5193a5a7c6ec09342b03b0f7da8373a15c51f36c2b9400175dc4bbec086dea"} Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.285266 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.286938 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.287554 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.290968 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.307424 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.328577 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.329007 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.329988 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330415 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blktr\" (UniqueName: \"kubernetes.io/projected/da5e5f63-b7aa-4cbb-89a3-584635543dec-kube-api-access-blktr\") pod \"dns-default-zjclt\" (UID: \"da5e5f63-b7aa-4cbb-89a3-584635543dec\") " pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330448 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6619f158-e866-4058-8719-d64ecc6b2858-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330482 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/228d7f4f-dc7b-4268-af44-a18948e63463-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2twbw\" (UID: \"228d7f4f-dc7b-4268-af44-a18948e63463\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330509 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/228d7f4f-dc7b-4268-af44-a18948e63463-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2twbw\" (UID: \"228d7f4f-dc7b-4268-af44-a18948e63463\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330553 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330583 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330611 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330644 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d52debbb-703b-4730-938f-493516301ac5-images\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330673 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p64rm\" (UniqueName: \"kubernetes.io/projected/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-kube-api-access-p64rm\") pod \"collect-profiles-29326230-67kmr\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330701 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28cb0326-cca7-4933-b96c-9293ef0ba9e2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-f6jqz\" (UID: \"28cb0326-cca7-4933-b96c-9293ef0ba9e2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330746 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a70e298-7f09-430f-8e74-2af8109b4885-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-95gbf\" (UID: \"7a70e298-7f09-430f-8e74-2af8109b4885\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330784 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0db13fcc-692c-451a-924b-4fc43cf14aea-cert\") pod \"ingress-canary-bn5rg\" (UID: \"0db13fcc-692c-451a-924b-4fc43cf14aea\") " pod="openshift-ingress-canary/ingress-canary-bn5rg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330810 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da5e5f63-b7aa-4cbb-89a3-584635543dec-config-volume\") pod \"dns-default-zjclt\" (UID: \"da5e5f63-b7aa-4cbb-89a3-584635543dec\") " pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330838 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99afa68f-99df-4f5f-a87c-562f49ba358e-serving-cert\") pod \"service-ca-operator-777779d784-hssvl\" (UID: \"99afa68f-99df-4f5f-a87c-562f49ba358e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330866 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j4lf\" (UniqueName: \"kubernetes.io/projected/d136451c-5940-45d2-8b32-0a8ea88cfdeb-kube-api-access-8j4lf\") pod \"machine-config-server-fzxjq\" (UID: \"d136451c-5940-45d2-8b32-0a8ea88cfdeb\") " pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330892 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sp4mz\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330920 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/05033876-b336-4e14-9dee-8761f4273ff0-stats-auth\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330954 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.330984 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxcns\" (UniqueName: \"kubernetes.io/projected/d52debbb-703b-4730-938f-493516301ac5-kube-api-access-gxcns\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.339049 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:53.839026251 +0000 UTC m=+142.263993131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.331007 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99afa68f-99df-4f5f-a87c-562f49ba358e-config\") pod \"service-ca-operator-777779d784-hssvl\" (UID: \"99afa68f-99df-4f5f-a87c-562f49ba358e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342506 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a70e298-7f09-430f-8e74-2af8109b4885-config\") pod \"kube-controller-manager-operator-78b949d7b-95gbf\" (UID: \"7a70e298-7f09-430f-8e74-2af8109b4885\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342562 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/acc502df-2c80-40fc-975f-d993333b1898-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342593 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-registry-certificates\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342622 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6vcs\" (UniqueName: \"kubernetes.io/projected/0db13fcc-692c-451a-924b-4fc43cf14aea-kube-api-access-d6vcs\") pod \"ingress-canary-bn5rg\" (UID: \"0db13fcc-692c-451a-924b-4fc43cf14aea\") " pod="openshift-ingress-canary/ingress-canary-bn5rg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342642 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-socket-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342675 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f690b48-b1ec-44c7-9371-8e2707cc7ef5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-k8qxt\" (UID: \"0f690b48-b1ec-44c7-9371-8e2707cc7ef5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342707 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a726141a-3bd2-41d5-9c11-003b460b0716-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wrf9q\" (UID: \"a726141a-3bd2-41d5-9c11-003b460b0716\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342737 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d136451c-5940-45d2-8b32-0a8ea88cfdeb-node-bootstrap-token\") pod \"machine-config-server-fzxjq\" (UID: \"d136451c-5940-45d2-8b32-0a8ea88cfdeb\") " pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342765 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/acc502df-2c80-40fc-975f-d993333b1898-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342796 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvm8p\" (UniqueName: \"kubernetes.io/projected/0f690b48-b1ec-44c7-9371-8e2707cc7ef5-kube-api-access-zvm8p\") pod \"kube-storage-version-migrator-operator-b67b599dd-k8qxt\" (UID: \"0f690b48-b1ec-44c7-9371-8e2707cc7ef5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342826 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342869 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ad583190-37cf-4e06-aef1-8ac44debcc16-proxy-tls\") pod \"machine-config-controller-84d6567774-vnq26\" (UID: \"ad583190-37cf-4e06-aef1-8ac44debcc16\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342896 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-mountpoint-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342923 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-registry-tls\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342949 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6tdc\" (UniqueName: \"kubernetes.io/projected/baa60884-bd0e-47b5-9471-ac2261b4be07-kube-api-access-r6tdc\") pod \"catalog-operator-68c6474976-w79hv\" (UID: \"baa60884-bd0e-47b5-9471-ac2261b4be07\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.342978 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h67tp\" (UniqueName: \"kubernetes.io/projected/4f79bb2f-4697-4f2f-80f5-c95f59b3273f-kube-api-access-h67tp\") pod \"control-plane-machine-set-operator-78cbb6b69f-5k929\" (UID: \"4f79bb2f-4697-4f2f-80f5-c95f59b3273f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343006 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-registration-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343052 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343081 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g7tl\" (UniqueName: \"kubernetes.io/projected/7a2c8998-2de1-453a-9adb-72facdc8bd2f-kube-api-access-9g7tl\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343106 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-plugins-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343176 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a726141a-3bd2-41d5-9c11-003b460b0716-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wrf9q\" (UID: \"a726141a-3bd2-41d5-9c11-003b460b0716\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343206 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pfnp\" (UniqueName: \"kubernetes.io/projected/4c52a35d-30a8-4ea9-964e-15f6a1658f1f-kube-api-access-7pfnp\") pod \"multus-admission-controller-857f4d67dd-gpd22\" (UID: \"4c52a35d-30a8-4ea9-964e-15f6a1658f1f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343232 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/baa60884-bd0e-47b5-9471-ac2261b4be07-srv-cert\") pod \"catalog-operator-68c6474976-w79hv\" (UID: \"baa60884-bd0e-47b5-9471-ac2261b4be07\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343253 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f42bf51c-38a4-4852-abbf-7ff85201f08d-webhook-cert\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343285 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2jwv\" (UniqueName: \"kubernetes.io/projected/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-kube-api-access-f2jwv\") pod \"marketplace-operator-79b997595-sp4mz\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343313 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6619f158-e866-4058-8719-d64ecc6b2858-trusted-ca\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343342 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgvfz\" (UniqueName: \"kubernetes.io/projected/792b69ef-6bb8-4fcd-8dc4-05d6105819d0-kube-api-access-vgvfz\") pod \"package-server-manager-789f6589d5-fphc8\" (UID: \"792b69ef-6bb8-4fcd-8dc4-05d6105819d0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343349 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343365 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll4lm\" (UniqueName: \"kubernetes.io/projected/ad583190-37cf-4e06-aef1-8ac44debcc16-kube-api-access-ll4lm\") pod \"machine-config-controller-84d6567774-vnq26\" (UID: \"ad583190-37cf-4e06-aef1-8ac44debcc16\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343424 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343455 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq8qh\" (UniqueName: \"kubernetes.io/projected/94ee183a-26ed-4067-bcae-719d049e6951-kube-api-access-nq8qh\") pod \"olm-operator-6b444d44fb-94l8v\" (UID: \"94ee183a-26ed-4067-bcae-719d049e6951\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343500 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-trusted-ca\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343533 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343572 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04031e34-327d-4a07-be9b-dd13473952cd-config\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343600 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/94ee183a-26ed-4067-bcae-719d049e6951-profile-collector-cert\") pod \"olm-operator-6b444d44fb-94l8v\" (UID: \"94ee183a-26ed-4067-bcae-719d049e6951\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343628 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05033876-b336-4e14-9dee-8761f4273ff0-service-ca-bundle\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343651 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/baa60884-bd0e-47b5-9471-ac2261b4be07-profile-collector-cert\") pod \"catalog-operator-68c6474976-w79hv\" (UID: \"baa60884-bd0e-47b5-9471-ac2261b4be07\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343679 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343707 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d52debbb-703b-4730-938f-493516301ac5-proxy-tls\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343732 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9c57d626-f993-4183-b903-fe1cc8c74daa-signing-key\") pod \"service-ca-9c57cc56f-fnszf\" (UID: \"9c57d626-f993-4183-b903-fe1cc8c74daa\") " pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343753 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sp4mz\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343815 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ad583190-37cf-4e06-aef1-8ac44debcc16-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vnq26\" (UID: \"ad583190-37cf-4e06-aef1-8ac44debcc16\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343838 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5vdh\" (UniqueName: \"kubernetes.io/projected/2f35d305-3b7d-4221-aa2d-3a2981ff6785-kube-api-access-p5vdh\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343882 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkjfv\" (UniqueName: \"kubernetes.io/projected/99afa68f-99df-4f5f-a87c-562f49ba358e-kube-api-access-tkjfv\") pod \"service-ca-operator-777779d784-hssvl\" (UID: \"99afa68f-99df-4f5f-a87c-562f49ba358e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343912 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/94ee183a-26ed-4067-bcae-719d049e6951-srv-cert\") pod \"olm-operator-6b444d44fb-94l8v\" (UID: \"94ee183a-26ed-4067-bcae-719d049e6951\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343937 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d136451c-5940-45d2-8b32-0a8ea88cfdeb-certs\") pod \"machine-config-server-fzxjq\" (UID: \"d136451c-5940-45d2-8b32-0a8ea88cfdeb\") " pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.343959 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04031e34-327d-4a07-be9b-dd13473952cd-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344033 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4f79bb2f-4697-4f2f-80f5-c95f59b3273f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5k929\" (UID: \"4f79bb2f-4697-4f2f-80f5-c95f59b3273f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344056 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-csi-data-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344081 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhznx\" (UniqueName: \"kubernetes.io/projected/05033876-b336-4e14-9dee-8761f4273ff0-kube-api-access-fhznx\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344106 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgzx6\" (UniqueName: \"kubernetes.io/projected/0893f102-0c49-4748-bf1d-1829c886acd4-kube-api-access-rgzx6\") pod \"migrator-59844c95c7-pmcrz\" (UID: \"0893f102-0c49-4748-bf1d-1829c886acd4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344131 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbkkv\" (UniqueName: \"kubernetes.io/projected/9c57d626-f993-4183-b903-fe1cc8c74daa-kube-api-access-zbkkv\") pod \"service-ca-9c57cc56f-fnszf\" (UID: \"9c57d626-f993-4183-b903-fe1cc8c74daa\") " pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344157 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f42bf51c-38a4-4852-abbf-7ff85201f08d-apiservice-cert\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344178 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-secret-volume\") pod \"collect-profiles-29326230-67kmr\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344207 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28cb0326-cca7-4933-b96c-9293ef0ba9e2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-f6jqz\" (UID: \"28cb0326-cca7-4933-b96c-9293ef0ba9e2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344246 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344273 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04031e34-327d-4a07-be9b-dd13473952cd-serving-cert\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344298 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f690b48-b1ec-44c7-9371-8e2707cc7ef5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-k8qxt\" (UID: \"0f690b48-b1ec-44c7-9371-8e2707cc7ef5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344324 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/05033876-b336-4e14-9dee-8761f4273ff0-default-certificate\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344354 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-policies\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344382 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d52debbb-703b-4730-938f-493516301ac5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344407 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28cb0326-cca7-4933-b96c-9293ef0ba9e2-config\") pod \"kube-apiserver-operator-766d6c64bb-f6jqz\" (UID: \"28cb0326-cca7-4933-b96c-9293ef0ba9e2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344432 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8hh6\" (UniqueName: \"kubernetes.io/projected/f42bf51c-38a4-4852-abbf-7ff85201f08d-kube-api-access-z8hh6\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344456 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/228d7f4f-dc7b-4268-af44-a18948e63463-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2twbw\" (UID: \"228d7f4f-dc7b-4268-af44-a18948e63463\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344483 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c52a35d-30a8-4ea9-964e-15f6a1658f1f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gpd22\" (UID: \"4c52a35d-30a8-4ea9-964e-15f6a1658f1f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344505 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-config-volume\") pod \"collect-profiles-29326230-67kmr\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344529 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-dir\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344557 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344601 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/da5e5f63-b7aa-4cbb-89a3-584635543dec-metrics-tls\") pod \"dns-default-zjclt\" (UID: \"da5e5f63-b7aa-4cbb-89a3-584635543dec\") " pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344626 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-bound-sa-token\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344651 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f42bf51c-38a4-4852-abbf-7ff85201f08d-tmpfs\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344677 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chvhg\" (UniqueName: \"kubernetes.io/projected/04031e34-327d-4a07-be9b-dd13473952cd-kube-api-access-chvhg\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344724 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltskc\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-kube-api-access-ltskc\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344753 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-542nt\" (UniqueName: \"kubernetes.io/projected/a726141a-3bd2-41d5-9c11-003b460b0716-kube-api-access-542nt\") pod \"openshift-controller-manager-operator-756b6f6bc6-wrf9q\" (UID: \"a726141a-3bd2-41d5-9c11-003b460b0716\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344780 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05033876-b336-4e14-9dee-8761f4273ff0-metrics-certs\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344807 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a70e298-7f09-430f-8e74-2af8109b4885-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-95gbf\" (UID: \"7a70e298-7f09-430f-8e74-2af8109b4885\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344833 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6619f158-e866-4058-8719-d64ecc6b2858-metrics-tls\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344860 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/792b69ef-6bb8-4fcd-8dc4-05d6105819d0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fphc8\" (UID: \"792b69ef-6bb8-4fcd-8dc4-05d6105819d0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344885 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2btbh\" (UniqueName: \"kubernetes.io/projected/6619f158-e866-4058-8719-d64ecc6b2858-kube-api-access-2btbh\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344914 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344940 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9c57d626-f993-4183-b903-fe1cc8c74daa-signing-cabundle\") pod \"service-ca-9c57cc56f-fnszf\" (UID: \"9c57d626-f993-4183-b903-fe1cc8c74daa\") " pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.344965 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04031e34-327d-4a07-be9b-dd13473952cd-service-ca-bundle\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.345761 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-mvbs7" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.346157 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.346694 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-policies\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.351420 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-trusted-ca\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.351723 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-dir\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.352585 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.353282 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.353527 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/acc502df-2c80-40fc-975f-d993333b1898-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.353967 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a726141a-3bd2-41d5-9c11-003b460b0716-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wrf9q\" (UID: \"a726141a-3bd2-41d5-9c11-003b460b0716\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.354567 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:53.854548954 +0000 UTC m=+142.279515834 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.355762 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.355993 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-registry-certificates\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.357130 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.358384 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.358557 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.359239 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.359944 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.361634 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.361676 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/acc502df-2c80-40fc-975f-d993333b1898-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.361942 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-registry-tls\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.362419 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.363814 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.364406 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a726141a-3bd2-41d5-9c11-003b460b0716-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wrf9q\" (UID: \"a726141a-3bd2-41d5-9c11-003b460b0716\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.366635 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.368286 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.376526 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.379844 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.435576 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltskc\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-kube-api-access-ltskc\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.442688 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-bound-sa-token\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.445570 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.445719 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:53.945688932 +0000 UTC m=+142.370655822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.445810 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq8qh\" (UniqueName: \"kubernetes.io/projected/94ee183a-26ed-4067-bcae-719d049e6951-kube-api-access-nq8qh\") pod \"olm-operator-6b444d44fb-94l8v\" (UID: \"94ee183a-26ed-4067-bcae-719d049e6951\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.445838 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.445864 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/baa60884-bd0e-47b5-9471-ac2261b4be07-profile-collector-cert\") pod \"catalog-operator-68c6474976-w79hv\" (UID: \"baa60884-bd0e-47b5-9471-ac2261b4be07\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.445889 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04031e34-327d-4a07-be9b-dd13473952cd-config\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.445912 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/94ee183a-26ed-4067-bcae-719d049e6951-profile-collector-cert\") pod \"olm-operator-6b444d44fb-94l8v\" (UID: \"94ee183a-26ed-4067-bcae-719d049e6951\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.445934 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05033876-b336-4e14-9dee-8761f4273ff0-service-ca-bundle\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.445955 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d52debbb-703b-4730-938f-493516301ac5-proxy-tls\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.445971 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9c57d626-f993-4183-b903-fe1cc8c74daa-signing-key\") pod \"service-ca-9c57cc56f-fnszf\" (UID: \"9c57d626-f993-4183-b903-fe1cc8c74daa\") " pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.445995 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sp4mz\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446104 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ad583190-37cf-4e06-aef1-8ac44debcc16-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vnq26\" (UID: \"ad583190-37cf-4e06-aef1-8ac44debcc16\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446125 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5vdh\" (UniqueName: \"kubernetes.io/projected/2f35d305-3b7d-4221-aa2d-3a2981ff6785-kube-api-access-p5vdh\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446141 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkjfv\" (UniqueName: \"kubernetes.io/projected/99afa68f-99df-4f5f-a87c-562f49ba358e-kube-api-access-tkjfv\") pod \"service-ca-operator-777779d784-hssvl\" (UID: \"99afa68f-99df-4f5f-a87c-562f49ba358e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446156 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04031e34-327d-4a07-be9b-dd13473952cd-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446171 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/94ee183a-26ed-4067-bcae-719d049e6951-srv-cert\") pod \"olm-operator-6b444d44fb-94l8v\" (UID: \"94ee183a-26ed-4067-bcae-719d049e6951\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446190 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d136451c-5940-45d2-8b32-0a8ea88cfdeb-certs\") pod \"machine-config-server-fzxjq\" (UID: \"d136451c-5940-45d2-8b32-0a8ea88cfdeb\") " pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446224 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4f79bb2f-4697-4f2f-80f5-c95f59b3273f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5k929\" (UID: \"4f79bb2f-4697-4f2f-80f5-c95f59b3273f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446247 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-csi-data-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446274 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhznx\" (UniqueName: \"kubernetes.io/projected/05033876-b336-4e14-9dee-8761f4273ff0-kube-api-access-fhznx\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446296 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgzx6\" (UniqueName: \"kubernetes.io/projected/0893f102-0c49-4748-bf1d-1829c886acd4-kube-api-access-rgzx6\") pod \"migrator-59844c95c7-pmcrz\" (UID: \"0893f102-0c49-4748-bf1d-1829c886acd4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446317 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbkkv\" (UniqueName: \"kubernetes.io/projected/9c57d626-f993-4183-b903-fe1cc8c74daa-kube-api-access-zbkkv\") pod \"service-ca-9c57cc56f-fnszf\" (UID: \"9c57d626-f993-4183-b903-fe1cc8c74daa\") " pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446340 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28cb0326-cca7-4933-b96c-9293ef0ba9e2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-f6jqz\" (UID: \"28cb0326-cca7-4933-b96c-9293ef0ba9e2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446361 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f42bf51c-38a4-4852-abbf-7ff85201f08d-apiservice-cert\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446379 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-secret-volume\") pod \"collect-profiles-29326230-67kmr\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446395 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04031e34-327d-4a07-be9b-dd13473952cd-serving-cert\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446411 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f690b48-b1ec-44c7-9371-8e2707cc7ef5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-k8qxt\" (UID: \"0f690b48-b1ec-44c7-9371-8e2707cc7ef5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446426 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/05033876-b336-4e14-9dee-8761f4273ff0-default-certificate\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446443 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d52debbb-703b-4730-938f-493516301ac5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446472 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28cb0326-cca7-4933-b96c-9293ef0ba9e2-config\") pod \"kube-apiserver-operator-766d6c64bb-f6jqz\" (UID: \"28cb0326-cca7-4933-b96c-9293ef0ba9e2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446489 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8hh6\" (UniqueName: \"kubernetes.io/projected/f42bf51c-38a4-4852-abbf-7ff85201f08d-kube-api-access-z8hh6\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446506 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/228d7f4f-dc7b-4268-af44-a18948e63463-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2twbw\" (UID: \"228d7f4f-dc7b-4268-af44-a18948e63463\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446524 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c52a35d-30a8-4ea9-964e-15f6a1658f1f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gpd22\" (UID: \"4c52a35d-30a8-4ea9-964e-15f6a1658f1f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446538 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-config-volume\") pod \"collect-profiles-29326230-67kmr\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446588 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/da5e5f63-b7aa-4cbb-89a3-584635543dec-metrics-tls\") pod \"dns-default-zjclt\" (UID: \"da5e5f63-b7aa-4cbb-89a3-584635543dec\") " pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446605 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f42bf51c-38a4-4852-abbf-7ff85201f08d-tmpfs\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446622 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chvhg\" (UniqueName: \"kubernetes.io/projected/04031e34-327d-4a07-be9b-dd13473952cd-kube-api-access-chvhg\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446653 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05033876-b336-4e14-9dee-8761f4273ff0-metrics-certs\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446668 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6619f158-e866-4058-8719-d64ecc6b2858-metrics-tls\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446685 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a70e298-7f09-430f-8e74-2af8109b4885-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-95gbf\" (UID: \"7a70e298-7f09-430f-8e74-2af8109b4885\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446701 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/792b69ef-6bb8-4fcd-8dc4-05d6105819d0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fphc8\" (UID: \"792b69ef-6bb8-4fcd-8dc4-05d6105819d0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446729 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2btbh\" (UniqueName: \"kubernetes.io/projected/6619f158-e866-4058-8719-d64ecc6b2858-kube-api-access-2btbh\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446745 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9c57d626-f993-4183-b903-fe1cc8c74daa-signing-cabundle\") pod \"service-ca-9c57cc56f-fnszf\" (UID: \"9c57d626-f993-4183-b903-fe1cc8c74daa\") " pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446760 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04031e34-327d-4a07-be9b-dd13473952cd-service-ca-bundle\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446777 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blktr\" (UniqueName: \"kubernetes.io/projected/da5e5f63-b7aa-4cbb-89a3-584635543dec-kube-api-access-blktr\") pod \"dns-default-zjclt\" (UID: \"da5e5f63-b7aa-4cbb-89a3-584635543dec\") " pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446792 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6619f158-e866-4058-8719-d64ecc6b2858-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446814 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/228d7f4f-dc7b-4268-af44-a18948e63463-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2twbw\" (UID: \"228d7f4f-dc7b-4268-af44-a18948e63463\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446828 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/228d7f4f-dc7b-4268-af44-a18948e63463-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2twbw\" (UID: \"228d7f4f-dc7b-4268-af44-a18948e63463\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446847 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p64rm\" (UniqueName: \"kubernetes.io/projected/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-kube-api-access-p64rm\") pod \"collect-profiles-29326230-67kmr\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446861 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d52debbb-703b-4730-938f-493516301ac5-images\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446884 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28cb0326-cca7-4933-b96c-9293ef0ba9e2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-f6jqz\" (UID: \"28cb0326-cca7-4933-b96c-9293ef0ba9e2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446900 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a70e298-7f09-430f-8e74-2af8109b4885-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-95gbf\" (UID: \"7a70e298-7f09-430f-8e74-2af8109b4885\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446922 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0db13fcc-692c-451a-924b-4fc43cf14aea-cert\") pod \"ingress-canary-bn5rg\" (UID: \"0db13fcc-692c-451a-924b-4fc43cf14aea\") " pod="openshift-ingress-canary/ingress-canary-bn5rg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446949 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da5e5f63-b7aa-4cbb-89a3-584635543dec-config-volume\") pod \"dns-default-zjclt\" (UID: \"da5e5f63-b7aa-4cbb-89a3-584635543dec\") " pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446963 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99afa68f-99df-4f5f-a87c-562f49ba358e-serving-cert\") pod \"service-ca-operator-777779d784-hssvl\" (UID: \"99afa68f-99df-4f5f-a87c-562f49ba358e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.446978 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j4lf\" (UniqueName: \"kubernetes.io/projected/d136451c-5940-45d2-8b32-0a8ea88cfdeb-kube-api-access-8j4lf\") pod \"machine-config-server-fzxjq\" (UID: \"d136451c-5940-45d2-8b32-0a8ea88cfdeb\") " pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447001 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/05033876-b336-4e14-9dee-8761f4273ff0-stats-auth\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447032 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sp4mz\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447058 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxcns\" (UniqueName: \"kubernetes.io/projected/d52debbb-703b-4730-938f-493516301ac5-kube-api-access-gxcns\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447076 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99afa68f-99df-4f5f-a87c-562f49ba358e-config\") pod \"service-ca-operator-777779d784-hssvl\" (UID: \"99afa68f-99df-4f5f-a87c-562f49ba358e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447107 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a70e298-7f09-430f-8e74-2af8109b4885-config\") pod \"kube-controller-manager-operator-78b949d7b-95gbf\" (UID: \"7a70e298-7f09-430f-8e74-2af8109b4885\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447162 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6vcs\" (UniqueName: \"kubernetes.io/projected/0db13fcc-692c-451a-924b-4fc43cf14aea-kube-api-access-d6vcs\") pod \"ingress-canary-bn5rg\" (UID: \"0db13fcc-692c-451a-924b-4fc43cf14aea\") " pod="openshift-ingress-canary/ingress-canary-bn5rg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447179 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-socket-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447204 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f690b48-b1ec-44c7-9371-8e2707cc7ef5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-k8qxt\" (UID: \"0f690b48-b1ec-44c7-9371-8e2707cc7ef5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447223 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d136451c-5940-45d2-8b32-0a8ea88cfdeb-node-bootstrap-token\") pod \"machine-config-server-fzxjq\" (UID: \"d136451c-5940-45d2-8b32-0a8ea88cfdeb\") " pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447240 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvm8p\" (UniqueName: \"kubernetes.io/projected/0f690b48-b1ec-44c7-9371-8e2707cc7ef5-kube-api-access-zvm8p\") pod \"kube-storage-version-migrator-operator-b67b599dd-k8qxt\" (UID: \"0f690b48-b1ec-44c7-9371-8e2707cc7ef5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447277 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ad583190-37cf-4e06-aef1-8ac44debcc16-proxy-tls\") pod \"machine-config-controller-84d6567774-vnq26\" (UID: \"ad583190-37cf-4e06-aef1-8ac44debcc16\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447295 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-mountpoint-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447312 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6tdc\" (UniqueName: \"kubernetes.io/projected/baa60884-bd0e-47b5-9471-ac2261b4be07-kube-api-access-r6tdc\") pod \"catalog-operator-68c6474976-w79hv\" (UID: \"baa60884-bd0e-47b5-9471-ac2261b4be07\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447327 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h67tp\" (UniqueName: \"kubernetes.io/projected/4f79bb2f-4697-4f2f-80f5-c95f59b3273f-kube-api-access-h67tp\") pod \"control-plane-machine-set-operator-78cbb6b69f-5k929\" (UID: \"4f79bb2f-4697-4f2f-80f5-c95f59b3273f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447342 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-registration-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447372 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-plugins-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447409 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pfnp\" (UniqueName: \"kubernetes.io/projected/4c52a35d-30a8-4ea9-964e-15f6a1658f1f-kube-api-access-7pfnp\") pod \"multus-admission-controller-857f4d67dd-gpd22\" (UID: \"4c52a35d-30a8-4ea9-964e-15f6a1658f1f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447424 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/baa60884-bd0e-47b5-9471-ac2261b4be07-srv-cert\") pod \"catalog-operator-68c6474976-w79hv\" (UID: \"baa60884-bd0e-47b5-9471-ac2261b4be07\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447438 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f42bf51c-38a4-4852-abbf-7ff85201f08d-webhook-cert\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447453 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6619f158-e866-4058-8719-d64ecc6b2858-trusted-ca\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447469 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2jwv\" (UniqueName: \"kubernetes.io/projected/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-kube-api-access-f2jwv\") pod \"marketplace-operator-79b997595-sp4mz\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447501 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgvfz\" (UniqueName: \"kubernetes.io/projected/792b69ef-6bb8-4fcd-8dc4-05d6105819d0-kube-api-access-vgvfz\") pod \"package-server-manager-789f6589d5-fphc8\" (UID: \"792b69ef-6bb8-4fcd-8dc4-05d6105819d0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.447516 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll4lm\" (UniqueName: \"kubernetes.io/projected/ad583190-37cf-4e06-aef1-8ac44debcc16-kube-api-access-ll4lm\") pod \"machine-config-controller-84d6567774-vnq26\" (UID: \"ad583190-37cf-4e06-aef1-8ac44debcc16\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.448135 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:53.948123786 +0000 UTC m=+142.373090766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.453570 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sp4mz\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.454694 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/05033876-b336-4e14-9dee-8761f4273ff0-service-ca-bundle\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.454784 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ad583190-37cf-4e06-aef1-8ac44debcc16-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-vnq26\" (UID: \"ad583190-37cf-4e06-aef1-8ac44debcc16\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.455366 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28cb0326-cca7-4933-b96c-9293ef0ba9e2-config\") pod \"kube-apiserver-operator-766d6c64bb-f6jqz\" (UID: \"28cb0326-cca7-4933-b96c-9293ef0ba9e2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.455791 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-registration-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.456131 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-plugins-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.456331 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/baa60884-bd0e-47b5-9471-ac2261b4be07-profile-collector-cert\") pod \"catalog-operator-68c6474976-w79hv\" (UID: \"baa60884-bd0e-47b5-9471-ac2261b4be07\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.457167 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/94ee183a-26ed-4067-bcae-719d049e6951-profile-collector-cert\") pod \"olm-operator-6b444d44fb-94l8v\" (UID: \"94ee183a-26ed-4067-bcae-719d049e6951\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.457764 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/9c57d626-f993-4183-b903-fe1cc8c74daa-signing-cabundle\") pod \"service-ca-9c57cc56f-fnszf\" (UID: \"9c57d626-f993-4183-b903-fe1cc8c74daa\") " pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.459116 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f42bf51c-38a4-4852-abbf-7ff85201f08d-apiservice-cert\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.459438 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d52debbb-703b-4730-938f-493516301ac5-proxy-tls\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.460545 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4f79bb2f-4697-4f2f-80f5-c95f59b3273f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-5k929\" (UID: \"4f79bb2f-4697-4f2f-80f5-c95f59b3273f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.460630 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-csi-data-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.462320 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f42bf51c-38a4-4852-abbf-7ff85201f08d-tmpfs\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.462470 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99afa68f-99df-4f5f-a87c-562f49ba358e-config\") pod \"service-ca-operator-777779d784-hssvl\" (UID: \"99afa68f-99df-4f5f-a87c-562f49ba358e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.462634 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sp4mz\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.463641 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04031e34-327d-4a07-be9b-dd13473952cd-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.464054 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-542nt\" (UniqueName: \"kubernetes.io/projected/a726141a-3bd2-41d5-9c11-003b460b0716-kube-api-access-542nt\") pod \"openshift-controller-manager-operator-756b6f6bc6-wrf9q\" (UID: \"a726141a-3bd2-41d5-9c11-003b460b0716\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.468876 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/da5e5f63-b7aa-4cbb-89a3-584635543dec-metrics-tls\") pod \"dns-default-zjclt\" (UID: \"da5e5f63-b7aa-4cbb-89a3-584635543dec\") " pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.469040 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-mountpoint-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.470180 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da5e5f63-b7aa-4cbb-89a3-584635543dec-config-volume\") pod \"dns-default-zjclt\" (UID: \"da5e5f63-b7aa-4cbb-89a3-584635543dec\") " pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.470462 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04031e34-327d-4a07-be9b-dd13473952cd-service-ca-bundle\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.470891 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/9c57d626-f993-4183-b903-fe1cc8c74daa-signing-key\") pod \"service-ca-9c57cc56f-fnszf\" (UID: \"9c57d626-f993-4183-b903-fe1cc8c74daa\") " pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.471859 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2f35d305-3b7d-4221-aa2d-3a2981ff6785-socket-dir\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.472594 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f690b48-b1ec-44c7-9371-8e2707cc7ef5-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-k8qxt\" (UID: \"0f690b48-b1ec-44c7-9371-8e2707cc7ef5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.475952 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-config-volume\") pod \"collect-profiles-29326230-67kmr\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.476310 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/228d7f4f-dc7b-4268-af44-a18948e63463-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2twbw\" (UID: \"228d7f4f-dc7b-4268-af44-a18948e63463\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.476470 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/baa60884-bd0e-47b5-9471-ac2261b4be07-srv-cert\") pod \"catalog-operator-68c6474976-w79hv\" (UID: \"baa60884-bd0e-47b5-9471-ac2261b4be07\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.476927 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d52debbb-703b-4730-938f-493516301ac5-images\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.477575 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ad583190-37cf-4e06-aef1-8ac44debcc16-proxy-tls\") pod \"machine-config-controller-84d6567774-vnq26\" (UID: \"ad583190-37cf-4e06-aef1-8ac44debcc16\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.477739 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6619f158-e866-4058-8719-d64ecc6b2858-metrics-tls\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.479541 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d52debbb-703b-4730-938f-493516301ac5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.480805 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f690b48-b1ec-44c7-9371-8e2707cc7ef5-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-k8qxt\" (UID: \"0f690b48-b1ec-44c7-9371-8e2707cc7ef5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.481547 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d136451c-5940-45d2-8b32-0a8ea88cfdeb-certs\") pod \"machine-config-server-fzxjq\" (UID: \"d136451c-5940-45d2-8b32-0a8ea88cfdeb\") " pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.482307 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a70e298-7f09-430f-8e74-2af8109b4885-config\") pod \"kube-controller-manager-operator-78b949d7b-95gbf\" (UID: \"7a70e298-7f09-430f-8e74-2af8109b4885\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.485353 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6619f158-e866-4058-8719-d64ecc6b2858-trusted-ca\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.491381 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04031e34-327d-4a07-be9b-dd13473952cd-config\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.492603 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/28cb0326-cca7-4933-b96c-9293ef0ba9e2-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-f6jqz\" (UID: \"28cb0326-cca7-4933-b96c-9293ef0ba9e2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.493174 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/05033876-b336-4e14-9dee-8761f4273ff0-metrics-certs\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.493318 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f42bf51c-38a4-4852-abbf-7ff85201f08d-webhook-cert\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.493913 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4c52a35d-30a8-4ea9-964e-15f6a1658f1f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-gpd22\" (UID: \"4c52a35d-30a8-4ea9-964e-15f6a1658f1f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.493985 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-secret-volume\") pod \"collect-profiles-29326230-67kmr\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.494356 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/05033876-b336-4e14-9dee-8761f4273ff0-stats-auth\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.496390 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99afa68f-99df-4f5f-a87c-562f49ba358e-serving-cert\") pod \"service-ca-operator-777779d784-hssvl\" (UID: \"99afa68f-99df-4f5f-a87c-562f49ba358e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.499644 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rxmh4"] Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.501895 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/792b69ef-6bb8-4fcd-8dc4-05d6105819d0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-fphc8\" (UID: \"792b69ef-6bb8-4fcd-8dc4-05d6105819d0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.502292 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/228d7f4f-dc7b-4268-af44-a18948e63463-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2twbw\" (UID: \"228d7f4f-dc7b-4268-af44-a18948e63463\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.506796 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d136451c-5940-45d2-8b32-0a8ea88cfdeb-node-bootstrap-token\") pod \"machine-config-server-fzxjq\" (UID: \"d136451c-5940-45d2-8b32-0a8ea88cfdeb\") " pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.507312 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0db13fcc-692c-451a-924b-4fc43cf14aea-cert\") pod \"ingress-canary-bn5rg\" (UID: \"0db13fcc-692c-451a-924b-4fc43cf14aea\") " pod="openshift-ingress-canary/ingress-canary-bn5rg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.508317 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/05033876-b336-4e14-9dee-8761f4273ff0-default-certificate\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.508393 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a70e298-7f09-430f-8e74-2af8109b4885-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-95gbf\" (UID: \"7a70e298-7f09-430f-8e74-2af8109b4885\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.511456 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04031e34-327d-4a07-be9b-dd13473952cd-serving-cert\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.527177 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/94ee183a-26ed-4067-bcae-719d049e6951-srv-cert\") pod \"olm-operator-6b444d44fb-94l8v\" (UID: \"94ee183a-26ed-4067-bcae-719d049e6951\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.527603 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g7tl\" (UniqueName: \"kubernetes.io/projected/7a2c8998-2de1-453a-9adb-72facdc8bd2f-kube-api-access-9g7tl\") pod \"oauth-openshift-558db77b4-bzljn\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.533543 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll4lm\" (UniqueName: \"kubernetes.io/projected/ad583190-37cf-4e06-aef1-8ac44debcc16-kube-api-access-ll4lm\") pod \"machine-config-controller-84d6567774-vnq26\" (UID: \"ad583190-37cf-4e06-aef1-8ac44debcc16\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.545040 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq8qh\" (UniqueName: \"kubernetes.io/projected/94ee183a-26ed-4067-bcae-719d049e6951-kube-api-access-nq8qh\") pod \"olm-operator-6b444d44fb-94l8v\" (UID: \"94ee183a-26ed-4067-bcae-719d049e6951\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.549591 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.550135 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.050117827 +0000 UTC m=+142.475084707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.568915 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkjfv\" (UniqueName: \"kubernetes.io/projected/99afa68f-99df-4f5f-a87c-562f49ba358e-kube-api-access-tkjfv\") pod \"service-ca-operator-777779d784-hssvl\" (UID: \"99afa68f-99df-4f5f-a87c-562f49ba358e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.587036 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5vdh\" (UniqueName: \"kubernetes.io/projected/2f35d305-3b7d-4221-aa2d-3a2981ff6785-kube-api-access-p5vdh\") pod \"csi-hostpathplugin-7mw6v\" (UID: \"2f35d305-3b7d-4221-aa2d-3a2981ff6785\") " pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.602288 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.615716 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h67tp\" (UniqueName: \"kubernetes.io/projected/4f79bb2f-4697-4f2f-80f5-c95f59b3273f-kube-api-access-h67tp\") pod \"control-plane-machine-set-operator-78cbb6b69f-5k929\" (UID: \"4f79bb2f-4697-4f2f-80f5-c95f59b3273f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.633300 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pfnp\" (UniqueName: \"kubernetes.io/projected/4c52a35d-30a8-4ea9-964e-15f6a1658f1f-kube-api-access-7pfnp\") pod \"multus-admission-controller-857f4d67dd-gpd22\" (UID: \"4c52a35d-30a8-4ea9-964e-15f6a1658f1f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.644315 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.654080 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.654486 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.154473569 +0000 UTC m=+142.579440449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.675224 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.688034 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.691188 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2jwv\" (UniqueName: \"kubernetes.io/projected/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-kube-api-access-f2jwv\") pod \"marketplace-operator-79b997595-sp4mz\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.703682 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8hh6\" (UniqueName: \"kubernetes.io/projected/f42bf51c-38a4-4852-abbf-7ff85201f08d-kube-api-access-z8hh6\") pod \"packageserver-d55dfcdfc-2jlq2\" (UID: \"f42bf51c-38a4-4852-abbf-7ff85201f08d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.708336 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgzx6\" (UniqueName: \"kubernetes.io/projected/0893f102-0c49-4748-bf1d-1829c886acd4-kube-api-access-rgzx6\") pod \"migrator-59844c95c7-pmcrz\" (UID: \"0893f102-0c49-4748-bf1d-1829c886acd4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.708703 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhznx\" (UniqueName: \"kubernetes.io/projected/05033876-b336-4e14-9dee-8761f4273ff0-kube-api-access-fhznx\") pod \"router-default-5444994796-gq2bg\" (UID: \"05033876-b336-4e14-9dee-8761f4273ff0\") " pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.727762 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8"] Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.727762 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbkkv\" (UniqueName: \"kubernetes.io/projected/9c57d626-f993-4183-b903-fe1cc8c74daa-kube-api-access-zbkkv\") pod \"service-ca-9c57cc56f-fnszf\" (UID: \"9c57d626-f993-4183-b903-fe1cc8c74daa\") " pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.742295 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.758326 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.758713 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.758840 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.25881036 +0000 UTC m=+142.683777240 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.758980 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.759357 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.259344359 +0000 UTC m=+142.684311239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.762200 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/28cb0326-cca7-4933-b96c-9293ef0ba9e2-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-f6jqz\" (UID: \"28cb0326-cca7-4933-b96c-9293ef0ba9e2\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.767662 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxcns\" (UniqueName: \"kubernetes.io/projected/d52debbb-703b-4730-938f-493516301ac5-kube-api-access-gxcns\") pod \"machine-config-operator-74547568cd-f8fgl\" (UID: \"d52debbb-703b-4730-938f-493516301ac5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.773259 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.789303 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.796606 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.804770 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chvhg\" (UniqueName: \"kubernetes.io/projected/04031e34-327d-4a07-be9b-dd13473952cd-kube-api-access-chvhg\") pod \"authentication-operator-69f744f599-2m4s6\" (UID: \"04031e34-327d-4a07-be9b-dd13473952cd\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.809685 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mlpcc"] Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.809889 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" Oct 04 10:36:53 crc kubenswrapper[5025]: W1004 10:36:53.815421 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod05033876_b336_4e14_9dee_8761f4273ff0.slice/crio-52b98d2dbeb7ac9143737cf0c93d158cbcc1ee1d6eef5b18b9b144939d778832 WatchSource:0}: Error finding container 52b98d2dbeb7ac9143737cf0c93d158cbcc1ee1d6eef5b18b9b144939d778832: Status 404 returned error can't find the container with id 52b98d2dbeb7ac9143737cf0c93d158cbcc1ee1d6eef5b18b9b144939d778832 Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.816917 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.827274 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.833997 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgvfz\" (UniqueName: \"kubernetes.io/projected/792b69ef-6bb8-4fcd-8dc4-05d6105819d0-kube-api-access-vgvfz\") pod \"package-server-manager-789f6589d5-fphc8\" (UID: \"792b69ef-6bb8-4fcd-8dc4-05d6105819d0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.841734 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6tdc\" (UniqueName: \"kubernetes.io/projected/baa60884-bd0e-47b5-9471-ac2261b4be07-kube-api-access-r6tdc\") pod \"catalog-operator-68c6474976-w79hv\" (UID: \"baa60884-bd0e-47b5-9471-ac2261b4be07\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.843943 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.847627 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j4lf\" (UniqueName: \"kubernetes.io/projected/d136451c-5940-45d2-8b32-0a8ea88cfdeb-kube-api-access-8j4lf\") pod \"machine-config-server-fzxjq\" (UID: \"d136451c-5940-45d2-8b32-0a8ea88cfdeb\") " pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.861311 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.861452 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.361428593 +0000 UTC m=+142.786395473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.861577 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.862093 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.362086255 +0000 UTC m=+142.787053135 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.862791 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6vcs\" (UniqueName: \"kubernetes.io/projected/0db13fcc-692c-451a-924b-4fc43cf14aea-kube-api-access-d6vcs\") pod \"ingress-canary-bn5rg\" (UID: \"0db13fcc-692c-451a-924b-4fc43cf14aea\") " pod="openshift-ingress-canary/ingress-canary-bn5rg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.881215 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.890625 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.919855 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/228d7f4f-dc7b-4268-af44-a18948e63463-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-2twbw\" (UID: \"228d7f4f-dc7b-4268-af44-a18948e63463\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.922004 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.922087 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blktr\" (UniqueName: \"kubernetes.io/projected/da5e5f63-b7aa-4cbb-89a3-584635543dec-kube-api-access-blktr\") pod \"dns-default-zjclt\" (UID: \"da5e5f63-b7aa-4cbb-89a3-584635543dec\") " pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.925719 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvm8p\" (UniqueName: \"kubernetes.io/projected/0f690b48-b1ec-44c7-9371-8e2707cc7ef5-kube-api-access-zvm8p\") pod \"kube-storage-version-migrator-operator-b67b599dd-k8qxt\" (UID: \"0f690b48-b1ec-44c7-9371-8e2707cc7ef5\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.960296 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-bn5rg" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.963476 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:53 crc kubenswrapper[5025]: E1004 10:36:53.963815 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.463800207 +0000 UTC m=+142.888767087 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.975305 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fzxjq" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.975633 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6619f158-e866-4058-8719-d64ecc6b2858-bound-sa-token\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.985293 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:53 crc kubenswrapper[5025]: I1004 10:36:53.997034 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.010537 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a70e298-7f09-430f-8e74-2af8109b4885-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-95gbf\" (UID: \"7a70e298-7f09-430f-8e74-2af8109b4885\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.011684 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p64rm\" (UniqueName: \"kubernetes.io/projected/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-kube-api-access-p64rm\") pod \"collect-profiles-29326230-67kmr\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.026663 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2btbh\" (UniqueName: \"kubernetes.io/projected/6619f158-e866-4058-8719-d64ecc6b2858-kube-api-access-2btbh\") pod \"ingress-operator-5b745b69d9-tzsdm\" (UID: \"6619f158-e866-4058-8719-d64ecc6b2858\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.032093 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.035430 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.057361 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.065354 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.065589 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:54 crc kubenswrapper[5025]: E1004 10:36:54.066082 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.566064656 +0000 UTC m=+142.991031536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.082455 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.164153 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.170258 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:54 crc kubenswrapper[5025]: E1004 10:36:54.170891 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.670876194 +0000 UTC m=+143.095843064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.221461 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.225215 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5bv89"] Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.269943 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s"] Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.271726 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:54 crc kubenswrapper[5025]: E1004 10:36:54.271996 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.771983714 +0000 UTC m=+143.196950594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.287658 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jbffs"] Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.305702 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6"] Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.310195 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v"] Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.330636 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z7zzf"] Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.371611 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gq2bg" event={"ID":"05033876-b336-4e14-9dee-8761f4273ff0","Type":"ContainerStarted","Data":"52b98d2dbeb7ac9143737cf0c93d158cbcc1ee1d6eef5b18b9b144939d778832"} Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.372371 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:54 crc kubenswrapper[5025]: E1004 10:36:54.372794 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.872776814 +0000 UTC m=+143.297743694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.376516 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" event={"ID":"003ddd1f-9614-417f-b8d9-e6efdc75b922","Type":"ContainerStarted","Data":"c4ad43e66e106ac61109591bf9dd27ed22076c40af4031ee2c54837564a0c02d"} Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.385044 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" event={"ID":"d92c5448-95d8-4008-a5a0-2c1a5a016db3","Type":"ContainerStarted","Data":"fa795aecb90344935c6264a60e2331d1e49cdf4487f83a5dcfb74b1575be2888"} Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.396613 5025 generic.go:334] "Generic (PLEG): container finished" podID="ed7f7173-bdd4-4ef2-a780-48c87a23125b" containerID="f5a0e25e4f479dae3988eef8f4d122f1bf749e474aa5d37a71f91109084fa6eb" exitCode=0 Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.400515 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" event={"ID":"ed7f7173-bdd4-4ef2-a780-48c87a23125b","Type":"ContainerDied","Data":"f5a0e25e4f479dae3988eef8f4d122f1bf749e474aa5d37a71f91109084fa6eb"} Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.400577 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" event={"ID":"ed7f7173-bdd4-4ef2-a780-48c87a23125b","Type":"ContainerStarted","Data":"b03a3c10c32bc2b04f149b146785387f9f831d5f545422f10ec132ef0a9bc6e6"} Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.474082 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:54 crc kubenswrapper[5025]: E1004 10:36:54.474421 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:54.974409483 +0000 UTC m=+143.399376363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.542418 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-mvbs7"] Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.542460 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl"] Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.542475 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz"] Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.574822 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:54 crc kubenswrapper[5025]: E1004 10:36:54.575144 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:55.07511771 +0000 UTC m=+143.500084590 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.689514 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:54 crc kubenswrapper[5025]: E1004 10:36:54.690331 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:55.190318704 +0000 UTC m=+143.615285584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.791565 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:54 crc kubenswrapper[5025]: E1004 10:36:54.792037 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:55.291999424 +0000 UTC m=+143.716966304 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.858768 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-t8lsk" podStartSLOduration=121.858748005 podStartE2EDuration="2m1.858748005s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:54.800874049 +0000 UTC m=+143.225840929" watchObservedRunningTime="2025-10-04 10:36:54.858748005 +0000 UTC m=+143.283714895" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.884766 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" podStartSLOduration=120.884745428 podStartE2EDuration="2m0.884745428s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:54.882750579 +0000 UTC m=+143.307717459" watchObservedRunningTime="2025-10-04 10:36:54.884745428 +0000 UTC m=+143.309712308" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.895412 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:54 crc kubenswrapper[5025]: E1004 10:36:54.895759 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:55.395748005 +0000 UTC m=+143.820714885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.955273 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-6mdn4" podStartSLOduration=120.955251388 podStartE2EDuration="2m0.955251388s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:54.926237892 +0000 UTC m=+143.351204782" watchObservedRunningTime="2025-10-04 10:36:54.955251388 +0000 UTC m=+143.380218268" Oct 04 10:36:54 crc kubenswrapper[5025]: I1004 10:36:54.996169 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:54 crc kubenswrapper[5025]: E1004 10:36:54.996615 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:55.496594477 +0000 UTC m=+143.921561357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.097456 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.097847 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:55.597827322 +0000 UTC m=+144.022794202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.201928 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.202043 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:55.702007638 +0000 UTC m=+144.126974518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.202925 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.203498 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:55.703483939 +0000 UTC m=+144.128450819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.303811 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.304244 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:55.804227127 +0000 UTC m=+144.229194007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.354257 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" podStartSLOduration=121.354201592 podStartE2EDuration="2m1.354201592s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:55.351449568 +0000 UTC m=+143.776416448" watchObservedRunningTime="2025-10-04 10:36:55.354201592 +0000 UTC m=+143.779168472" Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.411086 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.411398 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:55.911386515 +0000 UTC m=+144.336353395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.438213 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7mw6v"] Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.439806 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" event={"ID":"ed7f7173-bdd4-4ef2-a780-48c87a23125b","Type":"ContainerStarted","Data":"6feee7ea967daf5d09e01f1bf060c09e82d4a60aef73e11803d010452f916354"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.447104 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" event={"ID":"1613b097-9c6f-4ae1-84b3-fc12833ee019","Type":"ContainerStarted","Data":"928f1960e9e7f478b47e09b09d9f5d2cad4adcafef20d66ab6b20acaaf1fb73b"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.450743 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jbffs" event={"ID":"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120","Type":"ContainerStarted","Data":"81855c784b34b5f33172a675b123be4d7b7850c6d99b23097ea81151846dccf3"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.451768 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" event={"ID":"94ee183a-26ed-4067-bcae-719d049e6951","Type":"ContainerStarted","Data":"4e9d4a0a6ae94bf67977808f332cc3eeb3bc7f2f0d6583a541f18183c40ccdd9"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.454710 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" event={"ID":"dacae947-fe70-4eef-8f0c-8f012cea7c96","Type":"ContainerStarted","Data":"48d7fc9b1173f833d26df5ca6160cbd0e69f52c7e9e55395b049f00079a333fb"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.454768 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" event={"ID":"dacae947-fe70-4eef-8f0c-8f012cea7c96","Type":"ContainerStarted","Data":"0f68f85ecc76cd08ea93e2e51ff8f48890641fd3e48fa95c740f0a3fee8e031b"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.457180 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" event={"ID":"28cb0326-cca7-4933-b96c-9293ef0ba9e2","Type":"ContainerStarted","Data":"1f3260758a4119124863219aa39737f6118d363c0d5c3a072e22e5fdf84669d3"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.463901 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" event={"ID":"fb222f50-1834-4c9f-a6da-4d1bbd383398","Type":"ContainerStarted","Data":"843c6cf866e9d07b542ee51a5834e8ae360cc185095b8f742681cf544672f610"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.463960 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" event={"ID":"fb222f50-1834-4c9f-a6da-4d1bbd383398","Type":"ContainerStarted","Data":"ba450ede42fba9ffdd83cdb895673141b74ea2bea8d7b74bdfc77a130edfb413"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.486295 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" podStartSLOduration=122.486273965 podStartE2EDuration="2m2.486273965s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:55.470764673 +0000 UTC m=+143.895731583" watchObservedRunningTime="2025-10-04 10:36:55.486273965 +0000 UTC m=+143.911240845" Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.486703 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gq2bg" event={"ID":"05033876-b336-4e14-9dee-8761f4273ff0","Type":"ContainerStarted","Data":"bfe30afa688e2de4483233c34472baf5ab8e0d6853cee9bfaf0c857e228d318b"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.487263 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26"] Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.489026 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fzxjq" event={"ID":"d136451c-5940-45d2-8b32-0a8ea88cfdeb","Type":"ContainerStarted","Data":"7a1d67e62ded4a48d4f2a366b68c9d082abcc5247d6be3deac9e06f2ebdaf7b1"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.490090 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" event={"ID":"d52debbb-703b-4730-938f-493516301ac5","Type":"ContainerStarted","Data":"b94dbb5b8b2e89405bf40893e03c82717ef0822910caa01b8db412503f27c474"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.491460 5025 generic.go:334] "Generic (PLEG): container finished" podID="003ddd1f-9614-417f-b8d9-e6efdc75b922" containerID="2a6f06e2ac9486efd7c13de4ffc6679b0f6749e5680edc7dc1c2c869c40a89f8" exitCode=0 Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.491502 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" event={"ID":"003ddd1f-9614-417f-b8d9-e6efdc75b922","Type":"ContainerDied","Data":"2a6f06e2ac9486efd7c13de4ffc6679b0f6749e5680edc7dc1c2c869c40a89f8"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.493365 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" event={"ID":"d92c5448-95d8-4008-a5a0-2c1a5a016db3","Type":"ContainerStarted","Data":"e6ccc6e69caba70a67ec23aed0adb0026a6ed0eb760fddd219a0c19607bf4d21"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.512454 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.512613 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.012588359 +0000 UTC m=+144.437555239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.512862 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.513328 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5bv89" event={"ID":"0f85106f-c448-496a-8199-9a03990115ac","Type":"ContainerStarted","Data":"4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b"} Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.513335 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.013324384 +0000 UTC m=+144.438291264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.513375 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5bv89" event={"ID":"0f85106f-c448-496a-8199-9a03990115ac","Type":"ContainerStarted","Data":"0de25330acbe04083b83abfbb8f0d993ed949bbe0e4b4582627162b9c6e8053f"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.515161 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mvbs7" event={"ID":"f42407c5-6c5a-452c-8ad9-6ab593111376","Type":"ContainerStarted","Data":"f1e72db19d88daff267a169e85a804b871507bdf4de5f56f2e532ae1c9e73601"} Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.618363 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.618581 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.118533815 +0000 UTC m=+144.543500705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.618937 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.621729 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.121716845 +0000 UTC m=+144.546683735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.720485 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.720635 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.220615229 +0000 UTC m=+144.645582119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.720775 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.721123 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.221113516 +0000 UTC m=+144.646080396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.746298 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.767378 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:36:55 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:36:55 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:36:55 crc kubenswrapper[5025]: healthz check failed Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.767446 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.822551 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.822898 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.32288242 +0000 UTC m=+144.747849300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.840357 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpqmk" podStartSLOduration=122.840337569 podStartE2EDuration="2m2.840337569s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:55.779421578 +0000 UTC m=+144.204388478" watchObservedRunningTime="2025-10-04 10:36:55.840337569 +0000 UTC m=+144.265304449" Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.925351 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:55 crc kubenswrapper[5025]: E1004 10:36:55.936241 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.43622336 +0000 UTC m=+144.861190240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.938859 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv"] Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.946441 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-gpd22"] Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.952740 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-hssvl"] Oct 04 10:36:55 crc kubenswrapper[5025]: I1004 10:36:55.952794 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.039065 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.039175 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.539137843 +0000 UTC m=+144.964104723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.039787 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.040086 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.540061445 +0000 UTC m=+144.965028325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: W1004 10:36:56.054740 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda726141a_3bd2_41d5_9c11_003b460b0716.slice/crio-3c29489394b1ff656ab095abb92786fc1a3434c63a8a23ece86f448ef88d70a6 WatchSource:0}: Error finding container 3c29489394b1ff656ab095abb92786fc1a3434c63a8a23ece86f448ef88d70a6: Status 404 returned error can't find the container with id 3c29489394b1ff656ab095abb92786fc1a3434c63a8a23ece86f448ef88d70a6 Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.057508 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5bv89" podStartSLOduration=123.057493153 podStartE2EDuration="2m3.057493153s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.055254876 +0000 UTC m=+144.480221756" watchObservedRunningTime="2025-10-04 10:36:56.057493153 +0000 UTC m=+144.482460033" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.109518 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mlpcc" podStartSLOduration=123.109497058 podStartE2EDuration="2m3.109497058s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.104914081 +0000 UTC m=+144.529880961" watchObservedRunningTime="2025-10-04 10:36:56.109497058 +0000 UTC m=+144.534463938" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.152492 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.152950 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.154472 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.654451391 +0000 UTC m=+145.079418261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.157366 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-8kd6s" podStartSLOduration=122.15734961 podStartE2EDuration="2m2.15734961s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.152800464 +0000 UTC m=+144.577767344" watchObservedRunningTime="2025-10-04 10:36:56.15734961 +0000 UTC m=+144.582316490" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.181485 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.183328 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zcqx6" podStartSLOduration=122.183306551 podStartE2EDuration="2m2.183306551s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.177180921 +0000 UTC m=+144.602147801" watchObservedRunningTime="2025-10-04 10:36:56.183306551 +0000 UTC m=+144.608273431" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.227409 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-zjclt"] Oct 04 10:36:56 crc kubenswrapper[5025]: W1004 10:36:56.235128 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf42bf51c_38a4_4852_abbf_7ff85201f08d.slice/crio-ee04c183983dec5ab5230e4178c4fa9564991429c1cbe67de9bf86d1b65c5bb0 WatchSource:0}: Error finding container ee04c183983dec5ab5230e4178c4fa9564991429c1cbe67de9bf86d1b65c5bb0: Status 404 returned error can't find the container with id ee04c183983dec5ab5230e4178c4fa9564991429c1cbe67de9bf86d1b65c5bb0 Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.238549 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.244356 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fnszf"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.249491 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzljn"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.256761 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.257171 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.757159027 +0000 UTC m=+145.182125897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.268124 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sp4mz"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.289165 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-gq2bg" podStartSLOduration=122.289134174 podStartE2EDuration="2m2.289134174s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.258214743 +0000 UTC m=+144.683181623" watchObservedRunningTime="2025-10-04 10:36:56.289134174 +0000 UTC m=+144.714101054" Oct 04 10:36:56 crc kubenswrapper[5025]: W1004 10:36:56.294516 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda5e5f63_b7aa_4cbb_89a3_584635543dec.slice/crio-b7b2b7dd53ca26957a19fe0e357ab3871c4ac3caff362fe17857145ac8562166 WatchSource:0}: Error finding container b7b2b7dd53ca26957a19fe0e357ab3871c4ac3caff362fe17857145ac8562166: Status 404 returned error can't find the container with id b7b2b7dd53ca26957a19fe0e357ab3871c4ac3caff362fe17857145ac8562166 Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.318877 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.323092 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.326515 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-bn5rg"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.326875 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf"] Oct 04 10:36:56 crc kubenswrapper[5025]: W1004 10:36:56.329098 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c57d626_f993_4183_b903_fe1cc8c74daa.slice/crio-8edef6893489702ed58f422d997b24d96fbb25eaa50663da0730c9b7eecf9131 WatchSource:0}: Error finding container 8edef6893489702ed58f422d997b24d96fbb25eaa50663da0730c9b7eecf9131: Status 404 returned error can't find the container with id 8edef6893489702ed58f422d997b24d96fbb25eaa50663da0730c9b7eecf9131 Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.329560 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2m4s6"] Oct 04 10:36:56 crc kubenswrapper[5025]: W1004 10:36:56.334928 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a2c8998_2de1_453a_9adb_72facdc8bd2f.slice/crio-9766ce66c8de239c1c107e3f70622944739f26b464ca39b9b8da45007552fffe WatchSource:0}: Error finding container 9766ce66c8de239c1c107e3f70622944739f26b464ca39b9b8da45007552fffe: Status 404 returned error can't find the container with id 9766ce66c8de239c1c107e3f70622944739f26b464ca39b9b8da45007552fffe Oct 04 10:36:56 crc kubenswrapper[5025]: W1004 10:36:56.340631 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24cdcbf7_e6c5_4e50_b002_8beb6c1a41ef.slice/crio-f3b69e8ca74f9ec00e698996bef1d43fb052a92e8b820475b07e6340f5f40bcc WatchSource:0}: Error finding container f3b69e8ca74f9ec00e698996bef1d43fb052a92e8b820475b07e6340f5f40bcc: Status 404 returned error can't find the container with id f3b69e8ca74f9ec00e698996bef1d43fb052a92e8b820475b07e6340f5f40bcc Oct 04 10:36:56 crc kubenswrapper[5025]: W1004 10:36:56.352638 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a70e298_7f09_430f_8e74_2af8109b4885.slice/crio-38504d86708c7ce0561afa37f1e9d97907d587dde144cf86f2ed54b040aee514 WatchSource:0}: Error finding container 38504d86708c7ce0561afa37f1e9d97907d587dde144cf86f2ed54b040aee514: Status 404 returned error can't find the container with id 38504d86708c7ce0561afa37f1e9d97907d587dde144cf86f2ed54b040aee514 Oct 04 10:36:56 crc kubenswrapper[5025]: W1004 10:36:56.354500 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f690b48_b1ec_44c7_9371_8e2707cc7ef5.slice/crio-5ed672f0805dd26b58c35443adb7d8afc359720d9646c5b2a323a2591fb9a219 WatchSource:0}: Error finding container 5ed672f0805dd26b58c35443adb7d8afc359720d9646c5b2a323a2591fb9a219: Status 404 returned error can't find the container with id 5ed672f0805dd26b58c35443adb7d8afc359720d9646c5b2a323a2591fb9a219 Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.359686 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.360154 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.860135501 +0000 UTC m=+145.285102381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: W1004 10:36:56.380465 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04031e34_327d_4a07_be9b_dd13473952cd.slice/crio-3871bedd85a46403c9b44520cce8366cb043c23b6aa36a3d7820e50e060f1615 WatchSource:0}: Error finding container 3871bedd85a46403c9b44520cce8366cb043c23b6aa36a3d7820e50e060f1615: Status 404 returned error can't find the container with id 3871bedd85a46403c9b44520cce8366cb043c23b6aa36a3d7820e50e060f1615 Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.398299 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.398372 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.431790 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.461572 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.463671 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:56.963647854 +0000 UTC m=+145.388614734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.469218 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.484342 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.486065 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw"] Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.541676 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" event={"ID":"ad583190-37cf-4e06-aef1-8ac44debcc16","Type":"ContainerStarted","Data":"78e3e8e86f2f05ec91431404c7eb656ad335506a359396f22655bde99f75818e"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.541726 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" event={"ID":"ad583190-37cf-4e06-aef1-8ac44debcc16","Type":"ContainerStarted","Data":"0839e893b5835068a487aef1007548c2b051b83703186794e4c752a9b1e59aaa"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.545208 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz" event={"ID":"0893f102-0c49-4748-bf1d-1829c886acd4","Type":"ContainerStarted","Data":"3e71bfc425c43cd95a6adfd247c104c595efbfed46c9688c5043d2b72105363c"} Oct 04 10:36:56 crc kubenswrapper[5025]: W1004 10:36:56.549252 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod228d7f4f_dc7b_4268_af44_a18948e63463.slice/crio-79b31c4fe108a0ca985b4a772a52faf6fc39557cd272495a847fc70c5fc376f3 WatchSource:0}: Error finding container 79b31c4fe108a0ca985b4a772a52faf6fc39557cd272495a847fc70c5fc376f3: Status 404 returned error can't find the container with id 79b31c4fe108a0ca985b4a772a52faf6fc39557cd272495a847fc70c5fc376f3 Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.550349 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" event={"ID":"9c57d626-f993-4183-b903-fe1cc8c74daa","Type":"ContainerStarted","Data":"8edef6893489702ed58f422d997b24d96fbb25eaa50663da0730c9b7eecf9131"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.553709 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zjclt" event={"ID":"da5e5f63-b7aa-4cbb-89a3-584635543dec","Type":"ContainerStarted","Data":"b7b2b7dd53ca26957a19fe0e357ab3871c4ac3caff362fe17857145ac8562166"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.561715 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" event={"ID":"04031e34-327d-4a07-be9b-dd13473952cd","Type":"ContainerStarted","Data":"3871bedd85a46403c9b44520cce8366cb043c23b6aa36a3d7820e50e060f1615"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.562199 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.563190 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:57.0631633 +0000 UTC m=+145.488130190 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.576188 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" event={"ID":"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef","Type":"ContainerStarted","Data":"f3b69e8ca74f9ec00e698996bef1d43fb052a92e8b820475b07e6340f5f40bcc"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.577383 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" event={"ID":"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa","Type":"ContainerStarted","Data":"6fe9ec5428c260330d323256973017d25935bb4a2199e55b9561dd8cda613509"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.579694 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" event={"ID":"d52debbb-703b-4730-938f-493516301ac5","Type":"ContainerStarted","Data":"f7da505f106ff565b83b22379e9bfeef136377984cd89b452fd79472c3390473"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.587936 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jbffs" event={"ID":"ba4c3d7f-46dc-4262-9c8a-0b2991ce7120","Type":"ContainerStarted","Data":"891eeb905f47ddc72837b4a7c1b109389115053388adef97def879f89102edb7"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.588946 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.598224 5025 patch_prober.go:28] interesting pod/console-operator-58897d9998-jbffs container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/readyz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.598276 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jbffs" podUID="ba4c3d7f-46dc-4262-9c8a-0b2991ce7120" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/readyz\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.616399 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-jbffs" podStartSLOduration=123.616375847 podStartE2EDuration="2m3.616375847s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.612549125 +0000 UTC m=+145.037516005" watchObservedRunningTime="2025-10-04 10:36:56.616375847 +0000 UTC m=+145.041342727" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.618130 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" event={"ID":"7a2c8998-2de1-453a-9adb-72facdc8bd2f","Type":"ContainerStarted","Data":"9766ce66c8de239c1c107e3f70622944739f26b464ca39b9b8da45007552fffe"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.623490 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" event={"ID":"f42bf51c-38a4-4852-abbf-7ff85201f08d","Type":"ContainerStarted","Data":"ee04c183983dec5ab5230e4178c4fa9564991429c1cbe67de9bf86d1b65c5bb0"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.625142 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" event={"ID":"99afa68f-99df-4f5f-a87c-562f49ba358e","Type":"ContainerStarted","Data":"ee42e6e5bdbd2bfb552b02579411813772c14f3e9162d67e8bb5062e2fc99814"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.628623 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bn5rg" event={"ID":"0db13fcc-692c-451a-924b-4fc43cf14aea","Type":"ContainerStarted","Data":"754f4882e7a74f88b0ed756abc571d8e91681424697e2c94818f9d5d81a8d370"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.661853 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" podStartSLOduration=122.661834967 podStartE2EDuration="2m2.661834967s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.658276945 +0000 UTC m=+145.083243825" watchObservedRunningTime="2025-10-04 10:36:56.661834967 +0000 UTC m=+145.086801847" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.663913 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.665596 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:57.164416586 +0000 UTC m=+145.589383466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.679425 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fzxjq" event={"ID":"d136451c-5940-45d2-8b32-0a8ea88cfdeb","Type":"ContainerStarted","Data":"cbc63fb62d37f49e353860019d05473198b001c96398611e3742fc43416493ae"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.693736 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" event={"ID":"a726141a-3bd2-41d5-9c11-003b460b0716","Type":"ContainerStarted","Data":"c0755ab1acd2475baca26a4a79c8b9dc8350c47d1f622420b4e030c764d31d84"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.693777 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" event={"ID":"a726141a-3bd2-41d5-9c11-003b460b0716","Type":"ContainerStarted","Data":"3c29489394b1ff656ab095abb92786fc1a3434c63a8a23ece86f448ef88d70a6"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.695991 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" event={"ID":"7a70e298-7f09-430f-8e74-2af8109b4885","Type":"ContainerStarted","Data":"38504d86708c7ce0561afa37f1e9d97907d587dde144cf86f2ed54b040aee514"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.696880 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" event={"ID":"4f79bb2f-4697-4f2f-80f5-c95f59b3273f","Type":"ContainerStarted","Data":"72af4b6037e176a3dfdad22060e795393d3fadc89225a168be174f489adf63c9"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.697775 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" event={"ID":"28cb0326-cca7-4933-b96c-9293ef0ba9e2","Type":"ContainerStarted","Data":"6fa047ab1cbc27eaa5939790c65ca0b3dce90f3cc9b3b1645990b74500542300"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.699791 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" event={"ID":"1613b097-9c6f-4ae1-84b3-fc12833ee019","Type":"ContainerStarted","Data":"93fe97aa3cc95389c9312ccb91fc95d922f578de24b132bf14845a352dcd4ebb"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.701091 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" event={"ID":"4c52a35d-30a8-4ea9-964e-15f6a1658f1f","Type":"ContainerStarted","Data":"fa81f121221b120fcc6caec807f50acb67e6bbd3e376ce067dc6c73bdb9d5a6c"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.703778 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" event={"ID":"baa60884-bd0e-47b5-9471-ac2261b4be07","Type":"ContainerStarted","Data":"e8ccb0dd3fa3f5dc7ce25b2a2fd6364305c5fe6e463efea1eb1f79862259f7a5"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.704384 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.706092 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" event={"ID":"ed7f7173-bdd4-4ef2-a780-48c87a23125b","Type":"ContainerStarted","Data":"0347d032e5696db661948fec6a1436e9815696f51210c02802b0b0229131e46a"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.707563 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" event={"ID":"0f690b48-b1ec-44c7-9371-8e2707cc7ef5","Type":"ContainerStarted","Data":"5ed672f0805dd26b58c35443adb7d8afc359720d9646c5b2a323a2591fb9a219"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.723844 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" event={"ID":"792b69ef-6bb8-4fcd-8dc4-05d6105819d0","Type":"ContainerStarted","Data":"af7de332a1add414a275405918a4e3bb80d9c832d40f284b6349e48614feee5e"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.723984 5025 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w79hv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.724055 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" podUID="baa60884-bd0e-47b5-9471-ac2261b4be07" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.724855 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-fzxjq" podStartSLOduration=5.72484448 podStartE2EDuration="5.72484448s" podCreationTimestamp="2025-10-04 10:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.72105417 +0000 UTC m=+145.146021050" watchObservedRunningTime="2025-10-04 10:36:56.72484448 +0000 UTC m=+145.149811360" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.740187 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" event={"ID":"003ddd1f-9614-417f-b8d9-e6efdc75b922","Type":"ContainerStarted","Data":"9daec4fbcc15007a2c00b15206cda8c17d3f71f91706cb140f1f3bb84a80fb66"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.740941 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.743566 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" event={"ID":"94ee183a-26ed-4067-bcae-719d049e6951","Type":"ContainerStarted","Data":"3210beb8211d5440d1699c6765b382d98090711679cfdcaa857c3b0aa0832493"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.743923 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.751242 5025 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-94l8v container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.751297 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" podUID="94ee183a-26ed-4067-bcae-719d049e6951" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.752320 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:36:56 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:36:56 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:36:56 crc kubenswrapper[5025]: healthz check failed Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.752372 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.754449 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" event={"ID":"2f35d305-3b7d-4221-aa2d-3a2981ff6785","Type":"ContainerStarted","Data":"401bda520ad30ba0a774627b0b57c6e6a7a6b1e7ce57c8c82c12f850f095f875"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.765138 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.767420 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:57.267401041 +0000 UTC m=+145.692367921 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.773167 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-mvbs7" event={"ID":"f42407c5-6c5a-452c-8ad9-6ab593111376","Type":"ContainerStarted","Data":"281f49c0e788fa48f62ca7a30e3861bb52b6de49e50856af81769f2d791c11c4"} Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.773206 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-mvbs7" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.783403 5025 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvbs7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.783454 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mvbs7" podUID="f42407c5-6c5a-452c-8ad9-6ab593111376" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.790212 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qlhdg" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.792164 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" podStartSLOduration=122.79214261 podStartE2EDuration="2m2.79214261s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.750201671 +0000 UTC m=+145.175168551" watchObservedRunningTime="2025-10-04 10:36:56.79214261 +0000 UTC m=+145.217109490" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.792398 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-f6jqz" podStartSLOduration=122.792393639 podStartE2EDuration="2m2.792393639s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.770265249 +0000 UTC m=+145.195232129" watchObservedRunningTime="2025-10-04 10:36:56.792393639 +0000 UTC m=+145.217360519" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.813517 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wrf9q" podStartSLOduration=123.813496323 podStartE2EDuration="2m3.813496323s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.813352128 +0000 UTC m=+145.238319028" watchObservedRunningTime="2025-10-04 10:36:56.813496323 +0000 UTC m=+145.238463203" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.874085 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.878313 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:57.378300638 +0000 UTC m=+145.803267518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.928900 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" podStartSLOduration=122.928884234 podStartE2EDuration="2m2.928884234s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.863754268 +0000 UTC m=+145.288721148" watchObservedRunningTime="2025-10-04 10:36:56.928884234 +0000 UTC m=+145.353851114" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.978565 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:56 crc kubenswrapper[5025]: E1004 10:36:56.978913 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:57.4788754 +0000 UTC m=+145.903842280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.984337 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" podStartSLOduration=122.984322967 podStartE2EDuration="2m2.984322967s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.983406235 +0000 UTC m=+145.408373115" watchObservedRunningTime="2025-10-04 10:36:56.984322967 +0000 UTC m=+145.409289847" Oct 04 10:36:56 crc kubenswrapper[5025]: I1004 10:36:56.985194 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-mvbs7" podStartSLOduration=123.985188077 podStartE2EDuration="2m3.985188077s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:56.929381911 +0000 UTC m=+145.354348791" watchObservedRunningTime="2025-10-04 10:36:56.985188077 +0000 UTC m=+145.410154957" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.076465 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" podStartSLOduration=124.076450909 podStartE2EDuration="2m4.076450909s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:57.076398337 +0000 UTC m=+145.501365217" watchObservedRunningTime="2025-10-04 10:36:57.076450909 +0000 UTC m=+145.501417779" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.091085 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.091392 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:57.591380032 +0000 UTC m=+146.016346912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.209517 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.209860 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:57.709845678 +0000 UTC m=+146.134812558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.316229 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.316945 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:57.816932174 +0000 UTC m=+146.241899054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.417493 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.417731 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:57.917699763 +0000 UTC m=+146.342666663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.418005 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.419200 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:57.919188454 +0000 UTC m=+146.344155334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.519172 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.519530 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.019510887 +0000 UTC m=+146.444477777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.622827 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.623270 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.123256428 +0000 UTC m=+146.548223318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.724150 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.724388 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.224357328 +0000 UTC m=+146.649324208 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.724604 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.725025 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.22499248 +0000 UTC m=+146.649959360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.749678 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:36:57 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:36:57 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:36:57 crc kubenswrapper[5025]: healthz check failed Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.749772 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.778473 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" event={"ID":"4c52a35d-30a8-4ea9-964e-15f6a1658f1f","Type":"ContainerStarted","Data":"64ba40988f51fcd829e22d7513119c7a4d3c0b3edea421ebf202d9e23acee0bd"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.778526 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" event={"ID":"4c52a35d-30a8-4ea9-964e-15f6a1658f1f","Type":"ContainerStarted","Data":"df3910081de588b3204813fcaf3e6ab8d01924365b16fe20706aa0b3c8ae3449"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.779888 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-hssvl" event={"ID":"99afa68f-99df-4f5f-a87c-562f49ba358e","Type":"ContainerStarted","Data":"4329d9295b06936276c78c5a78125406420a5caa68012f83a767488f61577868"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.781293 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" event={"ID":"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa","Type":"ContainerStarted","Data":"326416e575dd0cdbb619a18fa4bb546727bf66104c9e149132bf3c2020c463cf"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.783562 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" event={"ID":"d52debbb-703b-4730-938f-493516301ac5","Type":"ContainerStarted","Data":"aec1420b1da76f4b9eaa140d49a612cd1ad2663f1140bd8a0713b8e36659c19f"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.785047 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" event={"ID":"7a70e298-7f09-430f-8e74-2af8109b4885","Type":"ContainerStarted","Data":"449a0cdd1ffa223182b59f61f22b04d494c418f4d8693b4834b3e5380fe788ae"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.786221 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" event={"ID":"2f35d305-3b7d-4221-aa2d-3a2981ff6785","Type":"ContainerStarted","Data":"089b4bc2aa9a392333b76e5b194f9f1c3aad9dc83c047ad4120f4f64d47c0236"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.787419 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" event={"ID":"4f79bb2f-4697-4f2f-80f5-c95f59b3273f","Type":"ContainerStarted","Data":"01db2b5b74f75b8d15df2de2225b915201545800814fc7dbd2570d2df0757f53"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.788886 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" event={"ID":"04031e34-327d-4a07-be9b-dd13473952cd","Type":"ContainerStarted","Data":"bd51bed5643f985c8a33d3362195eeb42a23764d033382de05cbd49064f7900f"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.790619 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" event={"ID":"7a2c8998-2de1-453a-9adb-72facdc8bd2f","Type":"ContainerStarted","Data":"b10e0caa164986d35464f6fe88fecb7214e3a05ad09a1201b0d82c82b98c8be3"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.791268 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.792805 5025 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bzljn container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" start-of-body= Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.792839 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" podUID="7a2c8998-2de1-453a-9adb-72facdc8bd2f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.793137 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz" event={"ID":"0893f102-0c49-4748-bf1d-1829c886acd4","Type":"ContainerStarted","Data":"502ac6391a536aeaddcd41cc799fbca3b35767e22d927c9868f2f0cda83fef7c"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.793160 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz" event={"ID":"0893f102-0c49-4748-bf1d-1829c886acd4","Type":"ContainerStarted","Data":"a1d95c9fa4dab6371bbdc1256d02cf176ea64a37c654b4745bf9645af96ba46d"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.794528 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" event={"ID":"baa60884-bd0e-47b5-9471-ac2261b4be07","Type":"ContainerStarted","Data":"80beeddb9d8d8d7fa8c97c7ad0cee3fb67fe29a95691545c1e4fd2ebbe422b3e"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.795164 5025 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-w79hv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.795193 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" podUID="baa60884-bd0e-47b5-9471-ac2261b4be07" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.796242 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" event={"ID":"228d7f4f-dc7b-4268-af44-a18948e63463","Type":"ContainerStarted","Data":"d177b2f25afaf240bd1d762df796768f2b1106a466e6556e7e3c98af10de48a8"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.796264 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" event={"ID":"228d7f4f-dc7b-4268-af44-a18948e63463","Type":"ContainerStarted","Data":"79b31c4fe108a0ca985b4a772a52faf6fc39557cd272495a847fc70c5fc376f3"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.798342 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" event={"ID":"6619f158-e866-4058-8719-d64ecc6b2858","Type":"ContainerStarted","Data":"a0c2d0e432a10fe57f2442c4dc26afe0ee3f1b821e10b1d4e9c02dbd86916a09"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.798363 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" event={"ID":"6619f158-e866-4058-8719-d64ecc6b2858","Type":"ContainerStarted","Data":"9f57fbdab670f6adbf7991e1f63c476c93916efe6c8af56d4732416f89d679c9"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.798372 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" event={"ID":"6619f158-e866-4058-8719-d64ecc6b2858","Type":"ContainerStarted","Data":"4a5061dfdadf901d8afae3be4f2ec663f402dea30b6e6d9942e5f39f0b10aa3e"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.800592 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" event={"ID":"1613b097-9c6f-4ae1-84b3-fc12833ee019","Type":"ContainerStarted","Data":"09114120d3216cfd3da6cca60fc7e1c673889713a215fccfaec62d26b5b80b6a"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.805254 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zjclt" event={"ID":"da5e5f63-b7aa-4cbb-89a3-584635543dec","Type":"ContainerStarted","Data":"20abd276efbecdaf19734bc2614aa30ac27b29be1de3a0fa18403cc76b22949c"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.805278 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-zjclt" event={"ID":"da5e5f63-b7aa-4cbb-89a3-584635543dec","Type":"ContainerStarted","Data":"3ad12b0d8255fd51eb07837fd173af66cd3f912976c56f8d3f31967eb39384a9"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.805389 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-zjclt" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.807839 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" event={"ID":"ad583190-37cf-4e06-aef1-8ac44debcc16","Type":"ContainerStarted","Data":"fffb99e1580a6b2a2364ea223cfbcf9f6d76469d65d04b4be942d2fe44236c5f"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.810081 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-bn5rg" event={"ID":"0db13fcc-692c-451a-924b-4fc43cf14aea","Type":"ContainerStarted","Data":"68188caf3e1221a95168c6264203caa917a548f78d775b94a591267b9ff5d8b3"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.811557 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" event={"ID":"0f690b48-b1ec-44c7-9371-8e2707cc7ef5","Type":"ContainerStarted","Data":"1f081ab0301e3d2620bb9ed611cb85950d44d38cac346b1b85c7fc57b4368b14"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.813242 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" event={"ID":"792b69ef-6bb8-4fcd-8dc4-05d6105819d0","Type":"ContainerStarted","Data":"1ff101e4b098ee47c8b527760b413c805f939096f41e1feafcb2fc59e24ef147"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.813264 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" event={"ID":"792b69ef-6bb8-4fcd-8dc4-05d6105819d0","Type":"ContainerStarted","Data":"e2a0df12fd574063da96ae619b3af8b491d4c473e580c6308fb25c8780ccc23b"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.813610 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.814922 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" event={"ID":"9c57d626-f993-4183-b903-fe1cc8c74daa","Type":"ContainerStarted","Data":"0d6b43e6f52a7234a22e20d61fff5794b337bcfb3efe3aacadff5a31b0356d1b"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.825819 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.826084 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.326069679 +0000 UTC m=+146.751036549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.826184 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.826447 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.326440832 +0000 UTC m=+146.751407712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.830207 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" event={"ID":"f42bf51c-38a4-4852-abbf-7ff85201f08d","Type":"ContainerStarted","Data":"fed5c43775a196569d7867693163b081e5e946c80b360f2f211d85594ea9ff0f"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.830879 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.833202 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" event={"ID":"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef","Type":"ContainerStarted","Data":"31be2214ed9ede5b72baef4d475d25d0c524b556cf3b5e322ddd4f38492570fb"} Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.833229 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.833581 5025 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvbs7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.833609 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mvbs7" podUID="f42407c5-6c5a-452c-8ad9-6ab593111376" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.833690 5025 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2jlq2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" start-of-body= Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.833741 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" podUID="f42bf51c-38a4-4852-abbf-7ff85201f08d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": dial tcp 10.217.0.41:5443: connect: connection refused" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.835584 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-95gbf" podStartSLOduration=123.835575606 podStartE2EDuration="2m3.835575606s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:57.835154971 +0000 UTC m=+146.260121851" watchObservedRunningTime="2025-10-04 10:36:57.835575606 +0000 UTC m=+146.260542486" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.837342 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-gpd22" podStartSLOduration=123.837334906 podStartE2EDuration="2m3.837334906s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:57.80365162 +0000 UTC m=+146.228618500" watchObservedRunningTime="2025-10-04 10:36:57.837334906 +0000 UTC m=+146.262301786" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.844105 5025 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sp4mz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.844159 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" podUID="24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.852639 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-94l8v" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.868943 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" podStartSLOduration=124.86892808 podStartE2EDuration="2m4.86892808s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:57.866541769 +0000 UTC m=+146.291508659" watchObservedRunningTime="2025-10-04 10:36:57.86892808 +0000 UTC m=+146.293894960" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.897261 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-tzsdm" podStartSLOduration=123.897248032 podStartE2EDuration="2m3.897248032s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:57.89658083 +0000 UTC m=+146.321547710" watchObservedRunningTime="2025-10-04 10:36:57.897248032 +0000 UTC m=+146.322214912" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.927620 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:57 crc kubenswrapper[5025]: E1004 10:36:57.929795 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.429776679 +0000 UTC m=+146.854743559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.930727 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" podStartSLOduration=123.930714191 podStartE2EDuration="2m3.930714191s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:57.929583172 +0000 UTC m=+146.354550052" watchObservedRunningTime="2025-10-04 10:36:57.930714191 +0000 UTC m=+146.355681071" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.980933 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-f8fgl" podStartSLOduration=123.980912584 podStartE2EDuration="2m3.980912584s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:57.979387702 +0000 UTC m=+146.404354582" watchObservedRunningTime="2025-10-04 10:36:57.980912584 +0000 UTC m=+146.405879464" Oct 04 10:36:57 crc kubenswrapper[5025]: I1004 10:36:57.981185 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-2twbw" podStartSLOduration=123.981179293 podStartE2EDuration="2m3.981179293s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:57.961907322 +0000 UTC m=+146.386874202" watchObservedRunningTime="2025-10-04 10:36:57.981179293 +0000 UTC m=+146.406146173" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.016266 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmcrz" podStartSLOduration=124.016248397 podStartE2EDuration="2m4.016248397s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.013448871 +0000 UTC m=+146.438415751" watchObservedRunningTime="2025-10-04 10:36:58.016248397 +0000 UTC m=+146.441215277" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.030982 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.031349 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.531326475 +0000 UTC m=+146.956293355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.068392 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-5k929" podStartSLOduration=124.068350976 podStartE2EDuration="2m4.068350976s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.04488648 +0000 UTC m=+146.469853360" watchObservedRunningTime="2025-10-04 10:36:58.068350976 +0000 UTC m=+146.493317856" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.103670 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2m4s6" podStartSLOduration=124.103632407 podStartE2EDuration="2m4.103632407s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.070162468 +0000 UTC m=+146.495129338" watchObservedRunningTime="2025-10-04 10:36:58.103632407 +0000 UTC m=+146.528599297" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.105628 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-z7zzf" podStartSLOduration=125.105621265 podStartE2EDuration="2m5.105621265s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.104873769 +0000 UTC m=+146.529840649" watchObservedRunningTime="2025-10-04 10:36:58.105621265 +0000 UTC m=+146.530588145" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.132096 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.132536 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.632519948 +0000 UTC m=+147.057486828 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.138659 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-zjclt" podStartSLOduration=7.138644079 podStartE2EDuration="7.138644079s" podCreationTimestamp="2025-10-04 10:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.135977187 +0000 UTC m=+146.560944057" watchObservedRunningTime="2025-10-04 10:36:58.138644079 +0000 UTC m=+146.563610959" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.214802 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-k8qxt" podStartSLOduration=124.214773752 podStartE2EDuration="2m4.214773752s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.168187553 +0000 UTC m=+146.593154433" watchObservedRunningTime="2025-10-04 10:36:58.214773752 +0000 UTC m=+146.639740632" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.222666 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.222747 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.235404 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.235803 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.735772973 +0000 UTC m=+147.160739853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.330398 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" podStartSLOduration=124.3303699 podStartE2EDuration="2m4.3303699s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.270476184 +0000 UTC m=+146.695443064" watchObservedRunningTime="2025-10-04 10:36:58.3303699 +0000 UTC m=+146.755336780" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.336670 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.337031 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.836999537 +0000 UTC m=+147.261966417 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.388768 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" podStartSLOduration=124.388702742 podStartE2EDuration="2m4.388702742s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.388600729 +0000 UTC m=+146.813567609" watchObservedRunningTime="2025-10-04 10:36:58.388702742 +0000 UTC m=+146.813669642" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.389288 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-bn5rg" podStartSLOduration=8.389283152 podStartE2EDuration="8.389283152s" podCreationTimestamp="2025-10-04 10:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.331419946 +0000 UTC m=+146.756386826" watchObservedRunningTime="2025-10-04 10:36:58.389283152 +0000 UTC m=+146.814250032" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.402336 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-jbffs" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.442080 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.442600 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:58.942579091 +0000 UTC m=+147.367545971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.455314 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" podStartSLOduration=124.455284907 podStartE2EDuration="2m4.455284907s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.45274141 +0000 UTC m=+146.877708290" watchObservedRunningTime="2025-10-04 10:36:58.455284907 +0000 UTC m=+146.880251787" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.511879 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-fnszf" podStartSLOduration=124.511849199 podStartE2EDuration="2m4.511849199s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.508807845 +0000 UTC m=+146.933774725" watchObservedRunningTime="2025-10-04 10:36:58.511849199 +0000 UTC m=+146.936816079" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.545264 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-vnq26" podStartSLOduration=124.545241135 podStartE2EDuration="2m4.545241135s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:58.53869046 +0000 UTC m=+146.963657340" watchObservedRunningTime="2025-10-04 10:36:58.545241135 +0000 UTC m=+146.970208005" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.547436 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.547634 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.047600056 +0000 UTC m=+147.472566936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.547852 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.548251 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.048232368 +0000 UTC m=+147.473199248 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.649789 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.649974 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.149946429 +0000 UTC m=+147.574913309 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.650183 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.650555 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.15054517 +0000 UTC m=+147.575512050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.751650 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.751753 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:36:58 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:36:58 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:36:58 crc kubenswrapper[5025]: healthz check failed Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.751812 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.752243 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.25222119 +0000 UTC m=+147.677188060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.837685 5025 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-sp4mz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.837733 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" podUID="24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.838545 5025 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-bzljn container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" start-of-body= Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.838644 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" podUID="7a2c8998-2de1-453a-9adb-72facdc8bd2f" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.16:6443/healthz\": dial tcp 10.217.0.16:6443: connect: connection refused" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.849067 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-w79hv" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.853185 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.853710 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.353695613 +0000 UTC m=+147.778662493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.949176 5025 patch_prober.go:28] interesting pod/apiserver-76f77b778f-rxmh4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]log ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]etcd ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]poststarthook/generic-apiserver-start-informers ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]poststarthook/max-in-flight-filter ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 04 10:36:58 crc kubenswrapper[5025]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 04 10:36:58 crc kubenswrapper[5025]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 04 10:36:58 crc kubenswrapper[5025]: [+]poststarthook/project.openshift.io-projectcache ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]poststarthook/openshift.io-startinformers ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 04 10:36:58 crc kubenswrapper[5025]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 04 10:36:58 crc kubenswrapper[5025]: livez check failed Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.949246 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" podUID="ed7f7173-bdd4-4ef2-a780-48c87a23125b" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:36:58 crc kubenswrapper[5025]: I1004 10:36:58.954138 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:58 crc kubenswrapper[5025]: E1004 10:36:58.955653 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.455625932 +0000 UTC m=+147.880592802 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.056433 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.056849 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.556832626 +0000 UTC m=+147.981799496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.157288 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.157407 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.657390688 +0000 UTC m=+148.082357568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.157500 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.157776 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.657769021 +0000 UTC m=+148.082735901 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.258806 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.258924 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.758900112 +0000 UTC m=+148.183866992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.259034 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.259385 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.759374788 +0000 UTC m=+148.184341668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.262450 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9bbj8" Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.359740 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.359923 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.859893719 +0000 UTC m=+148.284860599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.360128 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.360420 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.860407506 +0000 UTC m=+148.285374386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.461829 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.462028 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.961987853 +0000 UTC m=+148.386954733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.462290 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.462643 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:36:59.962635245 +0000 UTC m=+148.387602125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.563655 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.563841 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.063815828 +0000 UTC m=+148.488782708 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.564177 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.564516 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.064506152 +0000 UTC m=+148.489473032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.665498 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.665933 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.165919393 +0000 UTC m=+148.590886273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.747205 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:36:59 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:36:59 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:36:59 crc kubenswrapper[5025]: healthz check failed Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.747276 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.767359 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.768194 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.268160973 +0000 UTC m=+148.693127853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.777110 5025 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.838814 5025 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-2jlq2 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.838885 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" podUID="f42bf51c-38a4-4852-abbf-7ff85201f08d" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.41:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.856882 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" event={"ID":"2f35d305-3b7d-4221-aa2d-3a2981ff6785","Type":"ContainerStarted","Data":"67dbd8f6e42ef3d921f2064865932dc57ac23079cee3340168b1092feeef29a4"} Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.856929 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" event={"ID":"2f35d305-3b7d-4221-aa2d-3a2981ff6785","Type":"ContainerStarted","Data":"6c5df0354f5d3094421a5cf5afd5028f2630f2e595368378bd721c0489804b1a"} Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.856943 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" event={"ID":"2f35d305-3b7d-4221-aa2d-3a2981ff6785","Type":"ContainerStarted","Data":"385bf4ab613de3ba26ed068fcee204fac41fd7e354471e5068aa04f804515551"} Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.865678 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.868568 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.869182 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.369156009 +0000 UTC m=+148.794122889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.897838 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7mw6v" podStartSLOduration=9.897821663 podStartE2EDuration="9.897821663s" podCreationTimestamp="2025-10-04 10:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:36:59.895920208 +0000 UTC m=+148.320887088" watchObservedRunningTime="2025-10-04 10:36:59.897821663 +0000 UTC m=+148.322788543" Oct 04 10:36:59 crc kubenswrapper[5025]: I1004 10:36:59.970656 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:36:59 crc kubenswrapper[5025]: E1004 10:36:59.972134 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.472116754 +0000 UTC m=+148.897083634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.059529 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2jlq2" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.071623 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.071854 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.571825196 +0000 UTC m=+148.996792076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.071920 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.072226 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.572213309 +0000 UTC m=+148.997180179 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.173749 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.173949 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.673924121 +0000 UTC m=+149.098891001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.174278 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.174708 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.674690917 +0000 UTC m=+149.099657797 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.276220 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.276408 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.776382708 +0000 UTC m=+149.201349588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.276616 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.276713 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.276745 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.276829 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.277128 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.777109263 +0000 UTC m=+149.202076133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.277555 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.285874 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.297710 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.331454 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8k7kn"] Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.332722 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.335411 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.337399 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.369867 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8k7kn"] Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.371333 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.378576 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.378853 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.379235 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.879207727 +0000 UTC m=+149.304174597 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.385743 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.406651 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2j4wq"] Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.406862 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" podUID="df6dbb5c-fc0d-4020-af0e-f05bbd44949f" containerName="controller-manager" containerID="cri-o://618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c" gracePeriod=30 Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.480293 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-catalog-content\") pod \"certified-operators-8k7kn\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.480352 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdmnp\" (UniqueName: \"kubernetes.io/projected/a36bada3-22b5-48f9-b82e-26a30d24f57f-kube-api-access-bdmnp\") pod \"certified-operators-8k7kn\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.480372 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-utilities\") pod \"certified-operators-8k7kn\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.480418 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.480847 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-04 10:37:00.980830405 +0000 UTC m=+149.405797285 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvzjs" (UID: "acc502df-2c80-40fc-975f-d993333b1898") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.534095 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6xkcs"] Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.535038 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.545413 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.551684 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6xkcs"] Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.581180 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.581358 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-catalog-content\") pod \"certified-operators-8k7kn\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.581392 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdmnp\" (UniqueName: \"kubernetes.io/projected/a36bada3-22b5-48f9-b82e-26a30d24f57f-kube-api-access-bdmnp\") pod \"certified-operators-8k7kn\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.581415 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-utilities\") pod \"certified-operators-8k7kn\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.581973 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-utilities\") pod \"certified-operators-8k7kn\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.582100 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-04 10:37:01.082084871 +0000 UTC m=+149.507051751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.582310 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-catalog-content\") pod \"certified-operators-8k7kn\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.610529 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdmnp\" (UniqueName: \"kubernetes.io/projected/a36bada3-22b5-48f9-b82e-26a30d24f57f-kube-api-access-bdmnp\") pod \"certified-operators-8k7kn\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.619254 5025 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-04T10:36:59.777141351Z","Handler":null,"Name":""} Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.622861 5025 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.622901 5025 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.682864 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-catalog-content\") pod \"community-operators-6xkcs\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.682921 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.682976 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7qhg\" (UniqueName: \"kubernetes.io/projected/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-kube-api-access-c7qhg\") pod \"community-operators-6xkcs\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.683059 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-utilities\") pod \"community-operators-6xkcs\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.683113 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.685223 5025 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.685261 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.711714 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvzjs\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.713879 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.728127 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9l4lc"] Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.730079 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.737609 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9l4lc"] Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.747172 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:37:00 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:37:00 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:37:00 crc kubenswrapper[5025]: healthz check failed Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.747264 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:37:00 crc kubenswrapper[5025]: W1004 10:37:00.762538 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-1a29f4cafc6aee6b75243b30bcbbf8cdee2a89574f2b126cd640e542569a2733 WatchSource:0}: Error finding container 1a29f4cafc6aee6b75243b30bcbbf8cdee2a89574f2b126cd640e542569a2733: Status 404 returned error can't find the container with id 1a29f4cafc6aee6b75243b30bcbbf8cdee2a89574f2b126cd640e542569a2733 Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.784489 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.784648 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7qhg\" (UniqueName: \"kubernetes.io/projected/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-kube-api-access-c7qhg\") pod \"community-operators-6xkcs\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.784711 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-utilities\") pod \"community-operators-6xkcs\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.784733 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-catalog-content\") pod \"community-operators-6xkcs\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.785182 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-catalog-content\") pod \"community-operators-6xkcs\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.785716 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-utilities\") pod \"community-operators-6xkcs\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.804122 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.809878 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7qhg\" (UniqueName: \"kubernetes.io/projected/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-kube-api-access-c7qhg\") pod \"community-operators-6xkcs\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.866402 5025 generic.go:334] "Generic (PLEG): container finished" podID="cd17fb73-1ab6-46da-9b53-34ab62c9e0fa" containerID="326416e575dd0cdbb619a18fa4bb546727bf66104c9e149132bf3c2020c463cf" exitCode=0 Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.866467 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" event={"ID":"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa","Type":"ContainerDied","Data":"326416e575dd0cdbb619a18fa4bb546727bf66104c9e149132bf3c2020c463cf"} Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.870074 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"00c5971ad5d8dfc3399815ac1cd0dcd8b57984b6b96bab7d8b696dd4fa0c24d6"} Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.870998 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.871664 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"1a29f4cafc6aee6b75243b30bcbbf8cdee2a89574f2b126cd640e542569a2733"} Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.875215 5025 generic.go:334] "Generic (PLEG): container finished" podID="df6dbb5c-fc0d-4020-af0e-f05bbd44949f" containerID="618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c" exitCode=0 Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.875931 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.876046 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" event={"ID":"df6dbb5c-fc0d-4020-af0e-f05bbd44949f","Type":"ContainerDied","Data":"618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c"} Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.876066 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2j4wq" event={"ID":"df6dbb5c-fc0d-4020-af0e-f05bbd44949f","Type":"ContainerDied","Data":"26ee863bfc65c7286e87ffac56a29aeec821f27276962d950a9231777e521d37"} Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.876082 5025 scope.go:117] "RemoveContainer" containerID="618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.899085 5025 scope.go:117] "RemoveContainer" containerID="618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.899187 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2q4b\" (UniqueName: \"kubernetes.io/projected/a28fcea7-e325-4583-a1bb-64f7d80a00f7-kube-api-access-n2q4b\") pod \"certified-operators-9l4lc\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.899277 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-catalog-content\") pod \"certified-operators-9l4lc\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.899463 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-utilities\") pod \"certified-operators-9l4lc\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.900155 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c\": container with ID starting with 618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c not found: ID does not exist" containerID="618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.900256 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c"} err="failed to get container status \"618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c\": rpc error: code = NotFound desc = could not find container \"618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c\": container with ID starting with 618b30a274977600cd35f1e3978e6e792616088bba13bf0ea8eafb51c0e1119c not found: ID does not exist" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.900998 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.920117 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.933457 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n2pvj"] Oct 04 10:37:00 crc kubenswrapper[5025]: E1004 10:37:00.933873 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df6dbb5c-fc0d-4020-af0e-f05bbd44949f" containerName="controller-manager" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.933893 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="df6dbb5c-fc0d-4020-af0e-f05bbd44949f" containerName="controller-manager" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.934388 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="df6dbb5c-fc0d-4020-af0e-f05bbd44949f" containerName="controller-manager" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.937920 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:00 crc kubenswrapper[5025]: I1004 10:37:00.946642 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n2pvj"] Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.010054 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-client-ca\") pod \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.010143 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-config\") pod \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.010194 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8wvl\" (UniqueName: \"kubernetes.io/projected/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-kube-api-access-q8wvl\") pod \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.010220 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-serving-cert\") pod \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.010252 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-proxy-ca-bundles\") pod \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\" (UID: \"df6dbb5c-fc0d-4020-af0e-f05bbd44949f\") " Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.014082 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-utilities\") pod \"certified-operators-9l4lc\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.014505 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2q4b\" (UniqueName: \"kubernetes.io/projected/a28fcea7-e325-4583-a1bb-64f7d80a00f7-kube-api-access-n2q4b\") pod \"certified-operators-9l4lc\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.014633 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-catalog-content\") pod \"certified-operators-9l4lc\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.015603 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-client-ca" (OuterVolumeSpecName: "client-ca") pod "df6dbb5c-fc0d-4020-af0e-f05bbd44949f" (UID: "df6dbb5c-fc0d-4020-af0e-f05bbd44949f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.016881 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-utilities\") pod \"certified-operators-9l4lc\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.016995 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-config" (OuterVolumeSpecName: "config") pod "df6dbb5c-fc0d-4020-af0e-f05bbd44949f" (UID: "df6dbb5c-fc0d-4020-af0e-f05bbd44949f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.017900 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-catalog-content\") pod \"certified-operators-9l4lc\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.019113 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-w42lh"] Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.019843 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.023562 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "df6dbb5c-fc0d-4020-af0e-f05bbd44949f" (UID: "df6dbb5c-fc0d-4020-af0e-f05bbd44949f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.031567 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-kube-api-access-q8wvl" (OuterVolumeSpecName: "kube-api-access-q8wvl") pod "df6dbb5c-fc0d-4020-af0e-f05bbd44949f" (UID: "df6dbb5c-fc0d-4020-af0e-f05bbd44949f"). InnerVolumeSpecName "kube-api-access-q8wvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.040658 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2q4b\" (UniqueName: \"kubernetes.io/projected/a28fcea7-e325-4583-a1bb-64f7d80a00f7-kube-api-access-n2q4b\") pod \"certified-operators-9l4lc\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.041905 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "df6dbb5c-fc0d-4020-af0e-f05bbd44949f" (UID: "df6dbb5c-fc0d-4020-af0e-f05bbd44949f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.047107 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-w42lh"] Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.088524 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116294 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116356 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-utilities\") pod \"community-operators-n2pvj\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116384 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1877d566-a1e5-424c-87db-cb04d1fdc1b6-serving-cert\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116406 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-client-ca\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116479 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9d6z\" (UniqueName: \"kubernetes.io/projected/1877d566-a1e5-424c-87db-cb04d1fdc1b6-kube-api-access-s9d6z\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116505 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t6z5\" (UniqueName: \"kubernetes.io/projected/ef617b3d-a08d-4482-abd5-06a66cb36f45-kube-api-access-5t6z5\") pod \"community-operators-n2pvj\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116525 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-catalog-content\") pod \"community-operators-n2pvj\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116547 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-config\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116632 5025 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116646 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116658 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8wvl\" (UniqueName: \"kubernetes.io/projected/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-kube-api-access-q8wvl\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116669 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.116680 5025 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/df6dbb5c-fc0d-4020-af0e-f05bbd44949f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.119705 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8k7kn"] Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.218050 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9d6z\" (UniqueName: \"kubernetes.io/projected/1877d566-a1e5-424c-87db-cb04d1fdc1b6-kube-api-access-s9d6z\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.218323 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t6z5\" (UniqueName: \"kubernetes.io/projected/ef617b3d-a08d-4482-abd5-06a66cb36f45-kube-api-access-5t6z5\") pod \"community-operators-n2pvj\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.218340 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-catalog-content\") pod \"community-operators-n2pvj\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.218358 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-config\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.218411 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.218428 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-utilities\") pod \"community-operators-n2pvj\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.218447 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1877d566-a1e5-424c-87db-cb04d1fdc1b6-serving-cert\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.218466 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-client-ca\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.219356 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-client-ca\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.219999 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-catalog-content\") pod \"community-operators-n2pvj\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.220270 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-utilities\") pod \"community-operators-n2pvj\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.221449 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.221880 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-config\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.226885 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2j4wq"] Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.235245 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1877d566-a1e5-424c-87db-cb04d1fdc1b6-serving-cert\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.238619 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t6z5\" (UniqueName: \"kubernetes.io/projected/ef617b3d-a08d-4482-abd5-06a66cb36f45-kube-api-access-5t6z5\") pod \"community-operators-n2pvj\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.238864 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9d6z\" (UniqueName: \"kubernetes.io/projected/1877d566-a1e5-424c-87db-cb04d1fdc1b6-kube-api-access-s9d6z\") pod \"controller-manager-879f6c89f-w42lh\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.240651 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2j4wq"] Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.258157 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6xkcs"] Oct 04 10:37:01 crc kubenswrapper[5025]: W1004 10:37:01.264627 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded7e8716_6f98_43ad_aa7a_5a2b0b354091.slice/crio-d53c12b9d2eb10b19c9b6daa1a49f77d605c5554b9891c20ea37b9504912c7dc WatchSource:0}: Error finding container d53c12b9d2eb10b19c9b6daa1a49f77d605c5554b9891c20ea37b9504912c7dc: Status 404 returned error can't find the container with id d53c12b9d2eb10b19c9b6daa1a49f77d605c5554b9891c20ea37b9504912c7dc Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.268808 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.284148 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvzjs"] Oct 04 10:37:01 crc kubenswrapper[5025]: W1004 10:37:01.296938 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacc502df_2c80_40fc_975f_d993333b1898.slice/crio-f1e7fe83b58920d35473b32f01edf9e876904fe8bc4b9bdcd39d8f64d2235fcf WatchSource:0}: Error finding container f1e7fe83b58920d35473b32f01edf9e876904fe8bc4b9bdcd39d8f64d2235fcf: Status 404 returned error can't find the container with id f1e7fe83b58920d35473b32f01edf9e876904fe8bc4b9bdcd39d8f64d2235fcf Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.343075 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.358393 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9l4lc"] Oct 04 10:37:01 crc kubenswrapper[5025]: W1004 10:37:01.363118 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda28fcea7_e325_4583_a1bb_64f7d80a00f7.slice/crio-e18e15a00de83ed8e6a7884393d8576520ce2deaafd01c5777ac98d9631e6e73 WatchSource:0}: Error finding container e18e15a00de83ed8e6a7884393d8576520ce2deaafd01c5777ac98d9631e6e73: Status 404 returned error can't find the container with id e18e15a00de83ed8e6a7884393d8576520ce2deaafd01c5777ac98d9631e6e73 Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.525066 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n2pvj"] Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.600235 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-w42lh"] Oct 04 10:37:01 crc kubenswrapper[5025]: W1004 10:37:01.710130 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1877d566_a1e5_424c_87db_cb04d1fdc1b6.slice/crio-f37da1a433cfb303b3a65ddb705b304e04d2f5ba5fc6e5397cca454401467cab WatchSource:0}: Error finding container f37da1a433cfb303b3a65ddb705b304e04d2f5ba5fc6e5397cca454401467cab: Status 404 returned error can't find the container with id f37da1a433cfb303b3a65ddb705b304e04d2f5ba5fc6e5397cca454401467cab Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.751592 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:37:01 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:37:01 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:37:01 crc kubenswrapper[5025]: healthz check failed Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.752052 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.883435 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" event={"ID":"acc502df-2c80-40fc-975f-d993333b1898","Type":"ContainerStarted","Data":"e250e66e632078debc1e61557fcbcb5be7f1bcaa4dd4b24a3a499ffe69343f89"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.883544 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" event={"ID":"acc502df-2c80-40fc-975f-d993333b1898","Type":"ContainerStarted","Data":"f1e7fe83b58920d35473b32f01edf9e876904fe8bc4b9bdcd39d8f64d2235fcf"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.883616 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.885650 5025 generic.go:334] "Generic (PLEG): container finished" podID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerID="994bb912dd26aab48054dea9ca713672054ba76aaca69658676671c1d58a7f98" exitCode=0 Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.885773 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k7kn" event={"ID":"a36bada3-22b5-48f9-b82e-26a30d24f57f","Type":"ContainerDied","Data":"994bb912dd26aab48054dea9ca713672054ba76aaca69658676671c1d58a7f98"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.885835 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k7kn" event={"ID":"a36bada3-22b5-48f9-b82e-26a30d24f57f","Type":"ContainerStarted","Data":"2c6ded6cdd40b108760e5c77e027bce33aa708ddf83c0ec119751a8d02eec96d"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.887818 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"76494c1124ae47508fdc09c058da5bb55a6ae072799c9d391b69e370ce652a1d"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.888610 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.889768 5025 generic.go:334] "Generic (PLEG): container finished" podID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerID="0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f" exitCode=0 Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.889807 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9l4lc" event={"ID":"a28fcea7-e325-4583-a1bb-64f7d80a00f7","Type":"ContainerDied","Data":"0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.889845 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9l4lc" event={"ID":"a28fcea7-e325-4583-a1bb-64f7d80a00f7","Type":"ContainerStarted","Data":"e18e15a00de83ed8e6a7884393d8576520ce2deaafd01c5777ac98d9631e6e73"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.899724 5025 generic.go:334] "Generic (PLEG): container finished" podID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerID="9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075" exitCode=0 Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.900084 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pvj" event={"ID":"ef617b3d-a08d-4482-abd5-06a66cb36f45","Type":"ContainerDied","Data":"9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.900166 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pvj" event={"ID":"ef617b3d-a08d-4482-abd5-06a66cb36f45","Type":"ContainerStarted","Data":"093fdc339f8e39167f753e6da6556f00138e047b221818a28e309e5be8912710"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.903340 5025 generic.go:334] "Generic (PLEG): container finished" podID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerID="2391aab9f62481da328b20032accfb573c00df02062f8db7f650483b7d3d28c5" exitCode=0 Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.903558 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xkcs" event={"ID":"ed7e8716-6f98-43ad-aa7a-5a2b0b354091","Type":"ContainerDied","Data":"2391aab9f62481da328b20032accfb573c00df02062f8db7f650483b7d3d28c5"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.903630 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xkcs" event={"ID":"ed7e8716-6f98-43ad-aa7a-5a2b0b354091","Type":"ContainerStarted","Data":"d53c12b9d2eb10b19c9b6daa1a49f77d605c5554b9891c20ea37b9504912c7dc"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.906378 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" event={"ID":"1877d566-a1e5-424c-87db-cb04d1fdc1b6","Type":"ContainerStarted","Data":"7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.906408 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" event={"ID":"1877d566-a1e5-424c-87db-cb04d1fdc1b6","Type":"ContainerStarted","Data":"f37da1a433cfb303b3a65ddb705b304e04d2f5ba5fc6e5397cca454401467cab"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.906607 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.908438 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"58f797345ef80ccfbed49d72f1904b642375f0d75b0919a02bf200d526e88711"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.910122 5025 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-w42lh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" start-of-body= Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.910197 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" podUID="1877d566-a1e5-424c-87db-cb04d1fdc1b6" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.911414 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c1417dbcdaf6d167dda522e99f166d60de9e0c92876bceb1a6fd1491141905e4"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.911479 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d1e7da3467b9108fb9351b4f18b122421499737a954a910935260da49f3db288"} Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.912187 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:37:01 crc kubenswrapper[5025]: I1004 10:37:01.932301 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" podStartSLOduration=127.932277906 podStartE2EDuration="2m7.932277906s" podCreationTimestamp="2025-10-04 10:34:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:37:01.913876514 +0000 UTC m=+150.338843394" watchObservedRunningTime="2025-10-04 10:37:01.932277906 +0000 UTC m=+150.357244796" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.116969 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" podStartSLOduration=2.116950225 podStartE2EDuration="2.116950225s" podCreationTimestamp="2025-10-04 10:37:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:37:02.116896293 +0000 UTC m=+150.541863193" watchObservedRunningTime="2025-10-04 10:37:02.116950225 +0000 UTC m=+150.541917105" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.187008 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.362190 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p64rm\" (UniqueName: \"kubernetes.io/projected/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-kube-api-access-p64rm\") pod \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.362348 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-secret-volume\") pod \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.362384 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-config-volume\") pod \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\" (UID: \"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa\") " Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.363290 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-config-volume" (OuterVolumeSpecName: "config-volume") pod "cd17fb73-1ab6-46da-9b53-34ab62c9e0fa" (UID: "cd17fb73-1ab6-46da-9b53-34ab62c9e0fa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.372886 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cd17fb73-1ab6-46da-9b53-34ab62c9e0fa" (UID: "cd17fb73-1ab6-46da-9b53-34ab62c9e0fa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.379301 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-kube-api-access-p64rm" (OuterVolumeSpecName: "kube-api-access-p64rm") pod "cd17fb73-1ab6-46da-9b53-34ab62c9e0fa" (UID: "cd17fb73-1ab6-46da-9b53-34ab62c9e0fa"). InnerVolumeSpecName "kube-api-access-p64rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.418625 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.419172 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df6dbb5c-fc0d-4020-af0e-f05bbd44949f" path="/var/lib/kubelet/pods/df6dbb5c-fc0d-4020-af0e-f05bbd44949f/volumes" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.464429 5025 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.464511 5025 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.464537 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p64rm\" (UniqueName: \"kubernetes.io/projected/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa-kube-api-access-p64rm\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.523942 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p2fdt"] Oct 04 10:37:02 crc kubenswrapper[5025]: E1004 10:37:02.524196 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd17fb73-1ab6-46da-9b53-34ab62c9e0fa" containerName="collect-profiles" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.524208 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd17fb73-1ab6-46da-9b53-34ab62c9e0fa" containerName="collect-profiles" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.524308 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd17fb73-1ab6-46da-9b53-34ab62c9e0fa" containerName="collect-profiles" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.524979 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.528635 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.540366 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2fdt"] Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.666798 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-utilities\") pod \"redhat-marketplace-p2fdt\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.666891 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvdv2\" (UniqueName: \"kubernetes.io/projected/35966afc-f931-4779-a1bc-88b64819f756-kube-api-access-qvdv2\") pod \"redhat-marketplace-p2fdt\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.666928 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-catalog-content\") pod \"redhat-marketplace-p2fdt\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.746436 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:37:02 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:37:02 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:37:02 crc kubenswrapper[5025]: healthz check failed Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.746510 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.768312 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-utilities\") pod \"redhat-marketplace-p2fdt\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.768386 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvdv2\" (UniqueName: \"kubernetes.io/projected/35966afc-f931-4779-a1bc-88b64819f756-kube-api-access-qvdv2\") pod \"redhat-marketplace-p2fdt\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.768418 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-catalog-content\") pod \"redhat-marketplace-p2fdt\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.768974 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-catalog-content\") pod \"redhat-marketplace-p2fdt\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.769290 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-utilities\") pod \"redhat-marketplace-p2fdt\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.786618 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvdv2\" (UniqueName: \"kubernetes.io/projected/35966afc-f931-4779-a1bc-88b64819f756-kube-api-access-qvdv2\") pod \"redhat-marketplace-p2fdt\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.838350 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.934773 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x8lp9"] Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.942391 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.956105 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.956422 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr" event={"ID":"cd17fb73-1ab6-46da-9b53-34ab62c9e0fa","Type":"ContainerDied","Data":"6fe9ec5428c260330d323256973017d25935bb4a2199e55b9561dd8cda613509"} Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.956453 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fe9ec5428c260330d323256973017d25935bb4a2199e55b9561dd8cda613509" Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.968470 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x8lp9"] Oct 04 10:37:02 crc kubenswrapper[5025]: I1004 10:37:02.969798 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.072083 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-catalog-content\") pod \"redhat-marketplace-x8lp9\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.072161 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-utilities\") pod \"redhat-marketplace-x8lp9\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.072370 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npnfv\" (UniqueName: \"kubernetes.io/projected/37c2c485-27f8-40bf-b96f-20c99e41c354-kube-api-access-npnfv\") pod \"redhat-marketplace-x8lp9\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.173642 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npnfv\" (UniqueName: \"kubernetes.io/projected/37c2c485-27f8-40bf-b96f-20c99e41c354-kube-api-access-npnfv\") pod \"redhat-marketplace-x8lp9\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.173691 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-catalog-content\") pod \"redhat-marketplace-x8lp9\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.173712 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-utilities\") pod \"redhat-marketplace-x8lp9\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.174356 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-utilities\") pod \"redhat-marketplace-x8lp9\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.175728 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-catalog-content\") pod \"redhat-marketplace-x8lp9\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.221367 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npnfv\" (UniqueName: \"kubernetes.io/projected/37c2c485-27f8-40bf-b96f-20c99e41c354-kube-api-access-npnfv\") pod \"redhat-marketplace-x8lp9\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.234416 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.247573 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-rxmh4" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.269815 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.347359 5025 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvbs7 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.347411 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-mvbs7" podUID="f42407c5-6c5a-452c-8ad9-6ab593111376" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.347636 5025 patch_prober.go:28] interesting pod/downloads-7954f5f757-mvbs7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.347683 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-mvbs7" podUID="f42407c5-6c5a-452c-8ad9-6ab593111376" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.362403 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.362849 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.378165 5025 patch_prober.go:28] interesting pod/console-f9d7485db-5bv89 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.378226 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5bv89" podUID="0f85106f-c448-496a-8199-9a03990115ac" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.492759 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2fdt"] Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.525862 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xwws6"] Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.527312 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.540235 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.542936 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwws6"] Oct 04 10:37:03 crc kubenswrapper[5025]: W1004 10:37:03.580072 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35966afc_f931_4779_a1bc_88b64819f756.slice/crio-f56a9eacc2fe5811e570d27a04979fadfa9786ff4e0b2be38d5b2aa2f9059a00 WatchSource:0}: Error finding container f56a9eacc2fe5811e570d27a04979fadfa9786ff4e0b2be38d5b2aa2f9059a00: Status 404 returned error can't find the container with id f56a9eacc2fe5811e570d27a04979fadfa9786ff4e0b2be38d5b2aa2f9059a00 Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.687140 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-catalog-content\") pod \"redhat-operators-xwws6\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.687196 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwsfv\" (UniqueName: \"kubernetes.io/projected/84e2a31a-4153-4a4c-b56b-8480c8662e07-kube-api-access-nwsfv\") pod \"redhat-operators-xwws6\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.687252 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-utilities\") pod \"redhat-operators-xwws6\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.694666 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x8lp9"] Oct 04 10:37:03 crc kubenswrapper[5025]: W1004 10:37:03.721256 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37c2c485_27f8_40bf_b96f_20c99e41c354.slice/crio-1fed8fa1b6e8590c9e9beced446d606de7d5e1d769f50f460637c0b580bb44db WatchSource:0}: Error finding container 1fed8fa1b6e8590c9e9beced446d606de7d5e1d769f50f460637c0b580bb44db: Status 404 returned error can't find the container with id 1fed8fa1b6e8590c9e9beced446d606de7d5e1d769f50f460637c0b580bb44db Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.743424 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.752475 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:37:03 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:37:03 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:37:03 crc kubenswrapper[5025]: healthz check failed Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.752573 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.788591 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-catalog-content\") pod \"redhat-operators-xwws6\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.788630 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwsfv\" (UniqueName: \"kubernetes.io/projected/84e2a31a-4153-4a4c-b56b-8480c8662e07-kube-api-access-nwsfv\") pod \"redhat-operators-xwws6\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.788687 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-utilities\") pod \"redhat-operators-xwws6\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.789221 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-catalog-content\") pod \"redhat-operators-xwws6\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.789275 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-utilities\") pod \"redhat-operators-xwws6\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.808640 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwsfv\" (UniqueName: \"kubernetes.io/projected/84e2a31a-4153-4a4c-b56b-8480c8662e07-kube-api-access-nwsfv\") pod \"redhat-operators-xwws6\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.897133 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.903563 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.921443 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-74s25"] Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.923677 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.946195 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-74s25"] Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.967958 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8lp9" event={"ID":"37c2c485-27f8-40bf-b96f-20c99e41c354","Type":"ContainerStarted","Data":"1fed8fa1b6e8590c9e9beced446d606de7d5e1d769f50f460637c0b580bb44db"} Oct 04 10:37:03 crc kubenswrapper[5025]: I1004 10:37:03.969332 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2fdt" event={"ID":"35966afc-f931-4779-a1bc-88b64819f756","Type":"ContainerStarted","Data":"f56a9eacc2fe5811e570d27a04979fadfa9786ff4e0b2be38d5b2aa2f9059a00"} Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.093532 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6tgm\" (UniqueName: \"kubernetes.io/projected/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-kube-api-access-t6tgm\") pod \"redhat-operators-74s25\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.094328 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-catalog-content\") pod \"redhat-operators-74s25\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.094475 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-utilities\") pod \"redhat-operators-74s25\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.195723 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6tgm\" (UniqueName: \"kubernetes.io/projected/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-kube-api-access-t6tgm\") pod \"redhat-operators-74s25\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.196170 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-catalog-content\") pod \"redhat-operators-74s25\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.196199 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-utilities\") pod \"redhat-operators-74s25\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.196897 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-utilities\") pod \"redhat-operators-74s25\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.196902 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-catalog-content\") pod \"redhat-operators-74s25\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.212794 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6tgm\" (UniqueName: \"kubernetes.io/projected/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-kube-api-access-t6tgm\") pod \"redhat-operators-74s25\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.372387 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.444367 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.445047 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.445066 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwws6"] Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.445134 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.447103 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.447143 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 04 10:37:04 crc kubenswrapper[5025]: W1004 10:37:04.469206 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84e2a31a_4153_4a4c_b56b_8480c8662e07.slice/crio-6734cfda8d5f5d08f103501ae9fdda6f3074a159a868fea6366ea8fd304f7f10 WatchSource:0}: Error finding container 6734cfda8d5f5d08f103501ae9fdda6f3074a159a868fea6366ea8fd304f7f10: Status 404 returned error can't find the container with id 6734cfda8d5f5d08f103501ae9fdda6f3074a159a868fea6366ea8fd304f7f10 Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.532671 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.533359 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.539750 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.540151 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.540161 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.602786 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b83ab4d7-ed5d-4d43-b723-b1c401465fb0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.602896 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b83ab4d7-ed5d-4d43-b723-b1c401465fb0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.704526 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b83ab4d7-ed5d-4d43-b723-b1c401465fb0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.704945 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.704976 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.705037 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b83ab4d7-ed5d-4d43-b723-b1c401465fb0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.705106 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"b83ab4d7-ed5d-4d43-b723-b1c401465fb0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.739451 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"b83ab4d7-ed5d-4d43-b723-b1c401465fb0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.747964 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:37:04 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:37:04 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:37:04 crc kubenswrapper[5025]: healthz check failed Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.749234 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.787574 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-74s25"] Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.787867 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.806291 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.806341 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.806425 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.821546 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:37:04 crc kubenswrapper[5025]: I1004 10:37:04.873504 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.025234 5025 generic.go:334] "Generic (PLEG): container finished" podID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerID="431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582" exitCode=0 Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.025318 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8lp9" event={"ID":"37c2c485-27f8-40bf-b96f-20c99e41c354","Type":"ContainerDied","Data":"431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582"} Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.053375 5025 generic.go:334] "Generic (PLEG): container finished" podID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerID="ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736" exitCode=0 Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.053463 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwws6" event={"ID":"84e2a31a-4153-4a4c-b56b-8480c8662e07","Type":"ContainerDied","Data":"ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736"} Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.053489 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwws6" event={"ID":"84e2a31a-4153-4a4c-b56b-8480c8662e07","Type":"ContainerStarted","Data":"6734cfda8d5f5d08f103501ae9fdda6f3074a159a868fea6366ea8fd304f7f10"} Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.069007 5025 generic.go:334] "Generic (PLEG): container finished" podID="35966afc-f931-4779-a1bc-88b64819f756" containerID="3e8d7330cd9de49308d8f2da450b86f30fe2a44a32f9554d9899ddfd5b2440fc" exitCode=0 Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.069702 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2fdt" event={"ID":"35966afc-f931-4779-a1bc-88b64819f756","Type":"ContainerDied","Data":"3e8d7330cd9de49308d8f2da450b86f30fe2a44a32f9554d9899ddfd5b2440fc"} Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.079201 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74s25" event={"ID":"ec5571cf-ee2f-45e9-993b-85ec83a5e42f","Type":"ContainerStarted","Data":"000732ba3a400297bab93c8c84b01523fb8ad98a23e1b100da80476b15cc5fba"} Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.456315 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.726057 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 04 10:37:05 crc kubenswrapper[5025]: W1004 10:37:05.735106 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podef07b9a8_7ae9_4e6e_8189_bde3d75d7ddd.slice/crio-a73b93cc0e8d77d58acfc913d0a79ac6a1a92be14501f7b17b143222405f4132 WatchSource:0}: Error finding container a73b93cc0e8d77d58acfc913d0a79ac6a1a92be14501f7b17b143222405f4132: Status 404 returned error can't find the container with id a73b93cc0e8d77d58acfc913d0a79ac6a1a92be14501f7b17b143222405f4132 Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.748375 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:37:05 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:37:05 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:37:05 crc kubenswrapper[5025]: healthz check failed Oct 04 10:37:05 crc kubenswrapper[5025]: I1004 10:37:05.748464 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:37:06 crc kubenswrapper[5025]: I1004 10:37:06.097947 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd","Type":"ContainerStarted","Data":"a73b93cc0e8d77d58acfc913d0a79ac6a1a92be14501f7b17b143222405f4132"} Oct 04 10:37:06 crc kubenswrapper[5025]: I1004 10:37:06.101945 5025 generic.go:334] "Generic (PLEG): container finished" podID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerID="c871d6001409de5bddd83b8b4eb89083292e6fc16e0ea5e1832c0f806fad729b" exitCode=0 Oct 04 10:37:06 crc kubenswrapper[5025]: I1004 10:37:06.102260 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74s25" event={"ID":"ec5571cf-ee2f-45e9-993b-85ec83a5e42f","Type":"ContainerDied","Data":"c871d6001409de5bddd83b8b4eb89083292e6fc16e0ea5e1832c0f806fad729b"} Oct 04 10:37:06 crc kubenswrapper[5025]: I1004 10:37:06.111256 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b83ab4d7-ed5d-4d43-b723-b1c401465fb0","Type":"ContainerStarted","Data":"c7e7c413c6a76bccba2a810fa548a46eba135bf4e481f51ae2419f6e151f6458"} Oct 04 10:37:06 crc kubenswrapper[5025]: I1004 10:37:06.111294 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b83ab4d7-ed5d-4d43-b723-b1c401465fb0","Type":"ContainerStarted","Data":"fea16c5b17bc6d9a405f4594f160f4d9ef69b45e43658032287dc60f83bb0998"} Oct 04 10:37:06 crc kubenswrapper[5025]: I1004 10:37:06.750980 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:37:06 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:37:06 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:37:06 crc kubenswrapper[5025]: healthz check failed Oct 04 10:37:06 crc kubenswrapper[5025]: I1004 10:37:06.751343 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:37:07 crc kubenswrapper[5025]: I1004 10:37:07.125916 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd","Type":"ContainerStarted","Data":"fd9510b827233b0c0feb0ca6f0a17227d92c9b6265e5538d984b1fdebe27b8f2"} Oct 04 10:37:07 crc kubenswrapper[5025]: I1004 10:37:07.131860 5025 generic.go:334] "Generic (PLEG): container finished" podID="b83ab4d7-ed5d-4d43-b723-b1c401465fb0" containerID="c7e7c413c6a76bccba2a810fa548a46eba135bf4e481f51ae2419f6e151f6458" exitCode=0 Oct 04 10:37:07 crc kubenswrapper[5025]: I1004 10:37:07.131889 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b83ab4d7-ed5d-4d43-b723-b1c401465fb0","Type":"ContainerDied","Data":"c7e7c413c6a76bccba2a810fa548a46eba135bf4e481f51ae2419f6e151f6458"} Oct 04 10:37:07 crc kubenswrapper[5025]: I1004 10:37:07.142235 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.142220239 podStartE2EDuration="3.142220239s" podCreationTimestamp="2025-10-04 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:37:06.134753517 +0000 UTC m=+154.559720397" watchObservedRunningTime="2025-10-04 10:37:07.142220239 +0000 UTC m=+155.567187119" Oct 04 10:37:07 crc kubenswrapper[5025]: I1004 10:37:07.144920 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.144911442 podStartE2EDuration="3.144911442s" podCreationTimestamp="2025-10-04 10:37:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:37:07.141117491 +0000 UTC m=+155.566084371" watchObservedRunningTime="2025-10-04 10:37:07.144911442 +0000 UTC m=+155.569878322" Oct 04 10:37:07 crc kubenswrapper[5025]: I1004 10:37:07.747349 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:37:07 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:37:07 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:37:07 crc kubenswrapper[5025]: healthz check failed Oct 04 10:37:07 crc kubenswrapper[5025]: I1004 10:37:07.747451 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.142154 5025 generic.go:334] "Generic (PLEG): container finished" podID="ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd" containerID="fd9510b827233b0c0feb0ca6f0a17227d92c9b6265e5538d984b1fdebe27b8f2" exitCode=0 Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.142266 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd","Type":"ContainerDied","Data":"fd9510b827233b0c0feb0ca6f0a17227d92c9b6265e5538d984b1fdebe27b8f2"} Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.441082 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.611658 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kubelet-dir\") pod \"b83ab4d7-ed5d-4d43-b723-b1c401465fb0\" (UID: \"b83ab4d7-ed5d-4d43-b723-b1c401465fb0\") " Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.611720 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kube-api-access\") pod \"b83ab4d7-ed5d-4d43-b723-b1c401465fb0\" (UID: \"b83ab4d7-ed5d-4d43-b723-b1c401465fb0\") " Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.612861 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b83ab4d7-ed5d-4d43-b723-b1c401465fb0" (UID: "b83ab4d7-ed5d-4d43-b723-b1c401465fb0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.622634 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b83ab4d7-ed5d-4d43-b723-b1c401465fb0" (UID: "b83ab4d7-ed5d-4d43-b723-b1c401465fb0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.714984 5025 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.715032 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b83ab4d7-ed5d-4d43-b723-b1c401465fb0-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.748596 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:37:08 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:37:08 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:37:08 crc kubenswrapper[5025]: healthz check failed Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.748664 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:37:08 crc kubenswrapper[5025]: I1004 10:37:08.988773 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-zjclt" Oct 04 10:37:09 crc kubenswrapper[5025]: I1004 10:37:09.153274 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 04 10:37:09 crc kubenswrapper[5025]: I1004 10:37:09.153810 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"b83ab4d7-ed5d-4d43-b723-b1c401465fb0","Type":"ContainerDied","Data":"fea16c5b17bc6d9a405f4594f160f4d9ef69b45e43658032287dc60f83bb0998"} Oct 04 10:37:09 crc kubenswrapper[5025]: I1004 10:37:09.153874 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fea16c5b17bc6d9a405f4594f160f4d9ef69b45e43658032287dc60f83bb0998" Oct 04 10:37:09 crc kubenswrapper[5025]: I1004 10:37:09.745650 5025 patch_prober.go:28] interesting pod/router-default-5444994796-gq2bg container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 04 10:37:09 crc kubenswrapper[5025]: [-]has-synced failed: reason withheld Oct 04 10:37:09 crc kubenswrapper[5025]: [+]process-running ok Oct 04 10:37:09 crc kubenswrapper[5025]: healthz check failed Oct 04 10:37:09 crc kubenswrapper[5025]: I1004 10:37:09.746888 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gq2bg" podUID="05033876-b336-4e14-9dee-8761f4273ff0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 04 10:37:10 crc kubenswrapper[5025]: I1004 10:37:10.746353 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:37:10 crc kubenswrapper[5025]: I1004 10:37:10.751074 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-gq2bg" Oct 04 10:37:13 crc kubenswrapper[5025]: I1004 10:37:13.364297 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-mvbs7" Oct 04 10:37:13 crc kubenswrapper[5025]: I1004 10:37:13.389221 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:37:13 crc kubenswrapper[5025]: I1004 10:37:13.398438 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:37:14 crc kubenswrapper[5025]: I1004 10:37:14.714066 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:37:14 crc kubenswrapper[5025]: I1004 10:37:14.714153 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:37:15 crc kubenswrapper[5025]: I1004 10:37:15.623382 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:37:15 crc kubenswrapper[5025]: I1004 10:37:15.631411 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86312d38-63d0-409a-98d4-727f0fb47929-metrics-certs\") pod \"network-metrics-daemon-frc27\" (UID: \"86312d38-63d0-409a-98d4-727f0fb47929\") " pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:37:15 crc kubenswrapper[5025]: I1004 10:37:15.654282 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-frc27" Oct 04 10:37:20 crc kubenswrapper[5025]: I1004 10:37:20.926464 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:37:25 crc kubenswrapper[5025]: I1004 10:37:25.269583 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:37:25 crc kubenswrapper[5025]: I1004 10:37:25.282944 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd","Type":"ContainerDied","Data":"a73b93cc0e8d77d58acfc913d0a79ac6a1a92be14501f7b17b143222405f4132"} Oct 04 10:37:25 crc kubenswrapper[5025]: I1004 10:37:25.283059 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a73b93cc0e8d77d58acfc913d0a79ac6a1a92be14501f7b17b143222405f4132" Oct 04 10:37:25 crc kubenswrapper[5025]: I1004 10:37:25.283075 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 04 10:37:25 crc kubenswrapper[5025]: I1004 10:37:25.377386 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kube-api-access\") pod \"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd\" (UID: \"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd\") " Oct 04 10:37:25 crc kubenswrapper[5025]: I1004 10:37:25.377652 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kubelet-dir\") pod \"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd\" (UID: \"ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd\") " Oct 04 10:37:25 crc kubenswrapper[5025]: I1004 10:37:25.378155 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd" (UID: "ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:37:25 crc kubenswrapper[5025]: I1004 10:37:25.384852 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd" (UID: "ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:37:25 crc kubenswrapper[5025]: I1004 10:37:25.479657 5025 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:25 crc kubenswrapper[5025]: I1004 10:37:25.479692 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 04 10:37:32 crc kubenswrapper[5025]: E1004 10:37:32.198460 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 10:37:32 crc kubenswrapper[5025]: E1004 10:37:32.199113 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c7qhg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6xkcs_openshift-marketplace(ed7e8716-6f98-43ad-aa7a-5a2b0b354091): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:37:32 crc kubenswrapper[5025]: E1004 10:37:32.200290 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6xkcs" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" Oct 04 10:37:33 crc kubenswrapper[5025]: E1004 10:37:33.509209 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:b6f6b38447dd6e67a4ef5b2f35943cb7dd0448902ceb4ffe7efcf9b75a454fd8: Get \"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:b6f6b38447dd6e67a4ef5b2f35943cb7dd0448902ceb4ffe7efcf9b75a454fd8\": context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 10:37:33 crc kubenswrapper[5025]: E1004 10:37:33.509569 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t6tgm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-74s25_openshift-marketplace(ec5571cf-ee2f-45e9-993b-85ec83a5e42f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:b6f6b38447dd6e67a4ef5b2f35943cb7dd0448902ceb4ffe7efcf9b75a454fd8: Get \"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:b6f6b38447dd6e67a4ef5b2f35943cb7dd0448902ceb4ffe7efcf9b75a454fd8\": context canceled" logger="UnhandledError" Oct 04 10:37:33 crc kubenswrapper[5025]: E1004 10:37:33.509254 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6xkcs" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" Oct 04 10:37:33 crc kubenswrapper[5025]: E1004 10:37:33.510769 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:b6f6b38447dd6e67a4ef5b2f35943cb7dd0448902ceb4ffe7efcf9b75a454fd8: Get \\\"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:b6f6b38447dd6e67a4ef5b2f35943cb7dd0448902ceb4ffe7efcf9b75a454fd8\\\": context canceled\"" pod="openshift-marketplace/redhat-operators-74s25" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" Oct 04 10:37:33 crc kubenswrapper[5025]: E1004 10:37:33.592718 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 10:37:33 crc kubenswrapper[5025]: E1004 10:37:33.593298 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n2q4b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9l4lc_openshift-marketplace(a28fcea7-e325-4583-a1bb-64f7d80a00f7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:37:33 crc kubenswrapper[5025]: E1004 10:37:33.595212 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9l4lc" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" Oct 04 10:37:34 crc kubenswrapper[5025]: I1004 10:37:34.090721 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-fphc8" Oct 04 10:37:36 crc kubenswrapper[5025]: E1004 10:37:36.090251 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-74s25" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" Oct 04 10:37:36 crc kubenswrapper[5025]: E1004 10:37:36.090750 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9l4lc" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" Oct 04 10:37:38 crc kubenswrapper[5025]: E1004 10:37:38.692370 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 04 10:37:38 crc kubenswrapper[5025]: E1004 10:37:38.692877 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bdmnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8k7kn_openshift-marketplace(a36bada3-22b5-48f9-b82e-26a30d24f57f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:37:38 crc kubenswrapper[5025]: E1004 10:37:38.694686 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8k7kn" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" Oct 04 10:37:38 crc kubenswrapper[5025]: E1004 10:37:38.706283 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 04 10:37:38 crc kubenswrapper[5025]: E1004 10:37:38.706451 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5t6z5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-n2pvj_openshift-marketplace(ef617b3d-a08d-4482-abd5-06a66cb36f45): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:37:38 crc kubenswrapper[5025]: E1004 10:37:38.707634 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-n2pvj" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" Oct 04 10:37:38 crc kubenswrapper[5025]: E1004 10:37:38.711803 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 04 10:37:38 crc kubenswrapper[5025]: E1004 10:37:38.711917 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nwsfv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xwws6_openshift-marketplace(84e2a31a-4153-4a4c-b56b-8480c8662e07): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:37:38 crc kubenswrapper[5025]: E1004 10:37:38.713048 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xwws6" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" Oct 04 10:37:39 crc kubenswrapper[5025]: I1004 10:37:39.075743 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-frc27"] Oct 04 10:37:39 crc kubenswrapper[5025]: E1004 10:37:39.339663 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 10:37:39 crc kubenswrapper[5025]: E1004 10:37:39.339973 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-npnfv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-x8lp9_openshift-marketplace(37c2c485-27f8-40bf-b96f-20c99e41c354): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:37:39 crc kubenswrapper[5025]: E1004 10:37:39.341474 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-x8lp9" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" Oct 04 10:37:39 crc kubenswrapper[5025]: I1004 10:37:39.354613 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-frc27" event={"ID":"86312d38-63d0-409a-98d4-727f0fb47929","Type":"ContainerStarted","Data":"96c45dcef05809fcf96c4df6353ec414943ff802df1e01cd0e12441a9d1e630a"} Oct 04 10:37:39 crc kubenswrapper[5025]: E1004 10:37:39.357357 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-n2pvj" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" Oct 04 10:37:39 crc kubenswrapper[5025]: E1004 10:37:39.357660 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-x8lp9" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" Oct 04 10:37:39 crc kubenswrapper[5025]: E1004 10:37:39.357721 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xwws6" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" Oct 04 10:37:39 crc kubenswrapper[5025]: E1004 10:37:39.357776 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8k7kn" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" Oct 04 10:37:39 crc kubenswrapper[5025]: E1004 10:37:39.369484 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 04 10:37:39 crc kubenswrapper[5025]: E1004 10:37:39.369735 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qvdv2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-p2fdt_openshift-marketplace(35966afc-f931-4779-a1bc-88b64819f756): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 04 10:37:39 crc kubenswrapper[5025]: E1004 10:37:39.370948 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-p2fdt" podUID="35966afc-f931-4779-a1bc-88b64819f756" Oct 04 10:37:40 crc kubenswrapper[5025]: I1004 10:37:40.362981 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-frc27" event={"ID":"86312d38-63d0-409a-98d4-727f0fb47929","Type":"ContainerStarted","Data":"2e89b39f3c70e0ec66d589641c35f43a60c8102d20c093c2ca6b4015567e6f46"} Oct 04 10:37:40 crc kubenswrapper[5025]: I1004 10:37:40.363566 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-frc27" event={"ID":"86312d38-63d0-409a-98d4-727f0fb47929","Type":"ContainerStarted","Data":"0c2ea1006836762fe80b3cc320c9a76e5724d436afb21418dc05e7e8abb93006"} Oct 04 10:37:40 crc kubenswrapper[5025]: E1004 10:37:40.364959 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-p2fdt" podUID="35966afc-f931-4779-a1bc-88b64819f756" Oct 04 10:37:40 crc kubenswrapper[5025]: I1004 10:37:40.384856 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-frc27" podStartSLOduration=167.384835542 podStartE2EDuration="2m47.384835542s" podCreationTimestamp="2025-10-04 10:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:37:40.383785946 +0000 UTC m=+188.808752876" watchObservedRunningTime="2025-10-04 10:37:40.384835542 +0000 UTC m=+188.809802432" Oct 04 10:37:40 crc kubenswrapper[5025]: I1004 10:37:40.691219 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 04 10:37:44 crc kubenswrapper[5025]: I1004 10:37:44.714444 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:37:44 crc kubenswrapper[5025]: I1004 10:37:44.715282 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:37:47 crc kubenswrapper[5025]: I1004 10:37:47.412043 5025 generic.go:334] "Generic (PLEG): container finished" podID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerID="5ed121002a57ccad828494e90f66381f572564232e88cf63ccdc074b425fc83b" exitCode=0 Oct 04 10:37:47 crc kubenswrapper[5025]: I1004 10:37:47.412126 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xkcs" event={"ID":"ed7e8716-6f98-43ad-aa7a-5a2b0b354091","Type":"ContainerDied","Data":"5ed121002a57ccad828494e90f66381f572564232e88cf63ccdc074b425fc83b"} Oct 04 10:37:52 crc kubenswrapper[5025]: I1004 10:37:52.448703 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xkcs" event={"ID":"ed7e8716-6f98-43ad-aa7a-5a2b0b354091","Type":"ContainerStarted","Data":"5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809"} Oct 04 10:37:52 crc kubenswrapper[5025]: I1004 10:37:52.469214 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6xkcs" podStartSLOduration=3.491549209 podStartE2EDuration="52.469194715s" podCreationTimestamp="2025-10-04 10:37:00 +0000 UTC" firstStartedPulling="2025-10-04 10:37:01.90618116 +0000 UTC m=+150.331148060" lastFinishedPulling="2025-10-04 10:37:50.883826686 +0000 UTC m=+199.308793566" observedRunningTime="2025-10-04 10:37:52.467001304 +0000 UTC m=+200.891968194" watchObservedRunningTime="2025-10-04 10:37:52.469194715 +0000 UTC m=+200.894161595" Oct 04 10:37:56 crc kubenswrapper[5025]: I1004 10:37:56.489639 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8lp9" event={"ID":"37c2c485-27f8-40bf-b96f-20c99e41c354","Type":"ContainerDied","Data":"c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd"} Oct 04 10:37:56 crc kubenswrapper[5025]: I1004 10:37:56.489552 5025 generic.go:334] "Generic (PLEG): container finished" podID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerID="c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd" exitCode=0 Oct 04 10:37:56 crc kubenswrapper[5025]: I1004 10:37:56.495694 5025 generic.go:334] "Generic (PLEG): container finished" podID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerID="162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4" exitCode=0 Oct 04 10:37:56 crc kubenswrapper[5025]: I1004 10:37:56.495763 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9l4lc" event={"ID":"a28fcea7-e325-4583-a1bb-64f7d80a00f7","Type":"ContainerDied","Data":"162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4"} Oct 04 10:37:56 crc kubenswrapper[5025]: I1004 10:37:56.499948 5025 generic.go:334] "Generic (PLEG): container finished" podID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerID="f35f80e2c521d82a299254e2827250fb7fc9d36dcf99cbc52644263b1f059190" exitCode=0 Oct 04 10:37:56 crc kubenswrapper[5025]: I1004 10:37:56.500073 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74s25" event={"ID":"ec5571cf-ee2f-45e9-993b-85ec83a5e42f","Type":"ContainerDied","Data":"f35f80e2c521d82a299254e2827250fb7fc9d36dcf99cbc52644263b1f059190"} Oct 04 10:37:56 crc kubenswrapper[5025]: I1004 10:37:56.502666 5025 generic.go:334] "Generic (PLEG): container finished" podID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerID="3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874" exitCode=0 Oct 04 10:37:56 crc kubenswrapper[5025]: I1004 10:37:56.502687 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pvj" event={"ID":"ef617b3d-a08d-4482-abd5-06a66cb36f45","Type":"ContainerDied","Data":"3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874"} Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.509410 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8lp9" event={"ID":"37c2c485-27f8-40bf-b96f-20c99e41c354","Type":"ContainerStarted","Data":"20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e"} Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.511511 5025 generic.go:334] "Generic (PLEG): container finished" podID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerID="fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392" exitCode=0 Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.511562 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwws6" event={"ID":"84e2a31a-4153-4a4c-b56b-8480c8662e07","Type":"ContainerDied","Data":"fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392"} Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.516742 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9l4lc" event={"ID":"a28fcea7-e325-4583-a1bb-64f7d80a00f7","Type":"ContainerStarted","Data":"7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa"} Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.524053 5025 generic.go:334] "Generic (PLEG): container finished" podID="35966afc-f931-4779-a1bc-88b64819f756" containerID="a4d663f3225ebb067abd8feb86595c927c085e2192841892dfaf195a57992270" exitCode=0 Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.524141 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2fdt" event={"ID":"35966afc-f931-4779-a1bc-88b64819f756","Type":"ContainerDied","Data":"a4d663f3225ebb067abd8feb86595c927c085e2192841892dfaf195a57992270"} Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.527680 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74s25" event={"ID":"ec5571cf-ee2f-45e9-993b-85ec83a5e42f","Type":"ContainerStarted","Data":"843b0ec54670dd31f933f79691c32e2688b7f58d0d98a59f2a22220c8bb3b3b2"} Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.532880 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pvj" event={"ID":"ef617b3d-a08d-4482-abd5-06a66cb36f45","Type":"ContainerStarted","Data":"7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877"} Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.534905 5025 generic.go:334] "Generic (PLEG): container finished" podID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerID="01bfa63851deb6258320471d7ff712e86219ddff178d622f01280c0ec060acf0" exitCode=0 Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.534959 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k7kn" event={"ID":"a36bada3-22b5-48f9-b82e-26a30d24f57f","Type":"ContainerDied","Data":"01bfa63851deb6258320471d7ff712e86219ddff178d622f01280c0ec060acf0"} Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.536245 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x8lp9" podStartSLOduration=3.68075575 podStartE2EDuration="55.536230144s" podCreationTimestamp="2025-10-04 10:37:02 +0000 UTC" firstStartedPulling="2025-10-04 10:37:05.02808162 +0000 UTC m=+153.453048490" lastFinishedPulling="2025-10-04 10:37:56.883556004 +0000 UTC m=+205.308522884" observedRunningTime="2025-10-04 10:37:57.535040695 +0000 UTC m=+205.960007585" watchObservedRunningTime="2025-10-04 10:37:57.536230144 +0000 UTC m=+205.961197024" Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.570664 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-74s25" podStartSLOduration=3.739242563 podStartE2EDuration="54.570645075s" podCreationTimestamp="2025-10-04 10:37:03 +0000 UTC" firstStartedPulling="2025-10-04 10:37:06.103416652 +0000 UTC m=+154.528383532" lastFinishedPulling="2025-10-04 10:37:56.934819164 +0000 UTC m=+205.359786044" observedRunningTime="2025-10-04 10:37:57.570146899 +0000 UTC m=+205.995113779" watchObservedRunningTime="2025-10-04 10:37:57.570645075 +0000 UTC m=+205.995611955" Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.596235 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9l4lc" podStartSLOduration=2.520750114 podStartE2EDuration="57.596217788s" podCreationTimestamp="2025-10-04 10:37:00 +0000 UTC" firstStartedPulling="2025-10-04 10:37:01.892006564 +0000 UTC m=+150.316973474" lastFinishedPulling="2025-10-04 10:37:56.967474268 +0000 UTC m=+205.392441148" observedRunningTime="2025-10-04 10:37:57.594158101 +0000 UTC m=+206.019124981" watchObservedRunningTime="2025-10-04 10:37:57.596217788 +0000 UTC m=+206.021184668" Oct 04 10:37:57 crc kubenswrapper[5025]: I1004 10:37:57.637055 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n2pvj" podStartSLOduration=2.514649902 podStartE2EDuration="57.637037168s" podCreationTimestamp="2025-10-04 10:37:00 +0000 UTC" firstStartedPulling="2025-10-04 10:37:01.901489359 +0000 UTC m=+150.326456279" lastFinishedPulling="2025-10-04 10:37:57.023876665 +0000 UTC m=+205.448843545" observedRunningTime="2025-10-04 10:37:57.619114894 +0000 UTC m=+206.044081774" watchObservedRunningTime="2025-10-04 10:37:57.637037168 +0000 UTC m=+206.062004048" Oct 04 10:37:58 crc kubenswrapper[5025]: I1004 10:37:58.540837 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k7kn" event={"ID":"a36bada3-22b5-48f9-b82e-26a30d24f57f","Type":"ContainerStarted","Data":"25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f"} Oct 04 10:37:58 crc kubenswrapper[5025]: I1004 10:37:58.542688 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwws6" event={"ID":"84e2a31a-4153-4a4c-b56b-8480c8662e07","Type":"ContainerStarted","Data":"8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340"} Oct 04 10:37:58 crc kubenswrapper[5025]: I1004 10:37:58.544664 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2fdt" event={"ID":"35966afc-f931-4779-a1bc-88b64819f756","Type":"ContainerStarted","Data":"6fbe8530238f916aa361354131440486dbe0c47ca8c4fdb3a7e5a127017cc8b5"} Oct 04 10:37:58 crc kubenswrapper[5025]: I1004 10:37:58.560378 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8k7kn" podStartSLOduration=2.357366854 podStartE2EDuration="58.560363342s" podCreationTimestamp="2025-10-04 10:37:00 +0000 UTC" firstStartedPulling="2025-10-04 10:37:01.888317227 +0000 UTC m=+150.313284107" lastFinishedPulling="2025-10-04 10:37:58.091313715 +0000 UTC m=+206.516280595" observedRunningTime="2025-10-04 10:37:58.559316928 +0000 UTC m=+206.984283808" watchObservedRunningTime="2025-10-04 10:37:58.560363342 +0000 UTC m=+206.985330222" Oct 04 10:37:58 crc kubenswrapper[5025]: I1004 10:37:58.579364 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p2fdt" podStartSLOduration=3.751540559 podStartE2EDuration="56.579345401s" podCreationTimestamp="2025-10-04 10:37:02 +0000 UTC" firstStartedPulling="2025-10-04 10:37:05.07614357 +0000 UTC m=+153.501110440" lastFinishedPulling="2025-10-04 10:37:57.903948402 +0000 UTC m=+206.328915282" observedRunningTime="2025-10-04 10:37:58.575045471 +0000 UTC m=+207.000012351" watchObservedRunningTime="2025-10-04 10:37:58.579345401 +0000 UTC m=+207.004312281" Oct 04 10:37:58 crc kubenswrapper[5025]: I1004 10:37:58.598560 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xwws6" podStartSLOduration=2.672124416 podStartE2EDuration="55.598542486s" podCreationTimestamp="2025-10-04 10:37:03 +0000 UTC" firstStartedPulling="2025-10-04 10:37:05.055341706 +0000 UTC m=+153.480308586" lastFinishedPulling="2025-10-04 10:37:57.981759776 +0000 UTC m=+206.406726656" observedRunningTime="2025-10-04 10:37:58.594817155 +0000 UTC m=+207.019784055" watchObservedRunningTime="2025-10-04 10:37:58.598542486 +0000 UTC m=+207.023509366" Oct 04 10:38:00 crc kubenswrapper[5025]: I1004 10:38:00.715346 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:38:00 crc kubenswrapper[5025]: I1004 10:38:00.715848 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:38:00 crc kubenswrapper[5025]: I1004 10:38:00.902149 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:38:00 crc kubenswrapper[5025]: I1004 10:38:00.902589 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:38:01 crc kubenswrapper[5025]: I1004 10:38:01.020611 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:38:01 crc kubenswrapper[5025]: I1004 10:38:01.024525 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:38:01 crc kubenswrapper[5025]: I1004 10:38:01.090285 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:38:01 crc kubenswrapper[5025]: I1004 10:38:01.090347 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:38:01 crc kubenswrapper[5025]: I1004 10:38:01.137101 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:38:01 crc kubenswrapper[5025]: I1004 10:38:01.270066 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:38:01 crc kubenswrapper[5025]: I1004 10:38:01.270119 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:38:01 crc kubenswrapper[5025]: I1004 10:38:01.312364 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:38:01 crc kubenswrapper[5025]: I1004 10:38:01.604828 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:38:02 crc kubenswrapper[5025]: I1004 10:38:02.838555 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:38:02 crc kubenswrapper[5025]: I1004 10:38:02.838615 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:38:02 crc kubenswrapper[5025]: I1004 10:38:02.884291 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:38:03 crc kubenswrapper[5025]: I1004 10:38:03.271082 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:38:03 crc kubenswrapper[5025]: I1004 10:38:03.271149 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:38:03 crc kubenswrapper[5025]: I1004 10:38:03.321802 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:38:03 crc kubenswrapper[5025]: I1004 10:38:03.608846 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:38:03 crc kubenswrapper[5025]: I1004 10:38:03.613755 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:38:03 crc kubenswrapper[5025]: I1004 10:38:03.905050 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:38:03 crc kubenswrapper[5025]: I1004 10:38:03.905095 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:38:03 crc kubenswrapper[5025]: I1004 10:38:03.950799 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:38:04 crc kubenswrapper[5025]: I1004 10:38:04.372761 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:38:04 crc kubenswrapper[5025]: I1004 10:38:04.372849 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:38:04 crc kubenswrapper[5025]: I1004 10:38:04.421645 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:38:04 crc kubenswrapper[5025]: I1004 10:38:04.613901 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:38:04 crc kubenswrapper[5025]: I1004 10:38:04.626516 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:38:04 crc kubenswrapper[5025]: I1004 10:38:04.857289 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x8lp9"] Oct 04 10:38:05 crc kubenswrapper[5025]: I1004 10:38:05.579761 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x8lp9" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerName="registry-server" containerID="cri-o://20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e" gracePeriod=2 Oct 04 10:38:05 crc kubenswrapper[5025]: I1004 10:38:05.935694 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:38:05 crc kubenswrapper[5025]: I1004 10:38:05.985919 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-catalog-content\") pod \"37c2c485-27f8-40bf-b96f-20c99e41c354\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " Oct 04 10:38:05 crc kubenswrapper[5025]: I1004 10:38:05.986037 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-utilities\") pod \"37c2c485-27f8-40bf-b96f-20c99e41c354\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " Oct 04 10:38:05 crc kubenswrapper[5025]: I1004 10:38:05.986088 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npnfv\" (UniqueName: \"kubernetes.io/projected/37c2c485-27f8-40bf-b96f-20c99e41c354-kube-api-access-npnfv\") pod \"37c2c485-27f8-40bf-b96f-20c99e41c354\" (UID: \"37c2c485-27f8-40bf-b96f-20c99e41c354\") " Oct 04 10:38:05 crc kubenswrapper[5025]: I1004 10:38:05.986778 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-utilities" (OuterVolumeSpecName: "utilities") pod "37c2c485-27f8-40bf-b96f-20c99e41c354" (UID: "37c2c485-27f8-40bf-b96f-20c99e41c354"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:38:05 crc kubenswrapper[5025]: I1004 10:38:05.992794 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37c2c485-27f8-40bf-b96f-20c99e41c354-kube-api-access-npnfv" (OuterVolumeSpecName: "kube-api-access-npnfv") pod "37c2c485-27f8-40bf-b96f-20c99e41c354" (UID: "37c2c485-27f8-40bf-b96f-20c99e41c354"). InnerVolumeSpecName "kube-api-access-npnfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.012586 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37c2c485-27f8-40bf-b96f-20c99e41c354" (UID: "37c2c485-27f8-40bf-b96f-20c99e41c354"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.087155 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.087197 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npnfv\" (UniqueName: \"kubernetes.io/projected/37c2c485-27f8-40bf-b96f-20c99e41c354-kube-api-access-npnfv\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.087213 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37c2c485-27f8-40bf-b96f-20c99e41c354-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.587001 5025 generic.go:334] "Generic (PLEG): container finished" podID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerID="20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e" exitCode=0 Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.587088 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x8lp9" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.587115 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8lp9" event={"ID":"37c2c485-27f8-40bf-b96f-20c99e41c354","Type":"ContainerDied","Data":"20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e"} Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.587718 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x8lp9" event={"ID":"37c2c485-27f8-40bf-b96f-20c99e41c354","Type":"ContainerDied","Data":"1fed8fa1b6e8590c9e9beced446d606de7d5e1d769f50f460637c0b580bb44db"} Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.587744 5025 scope.go:117] "RemoveContainer" containerID="20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.612509 5025 scope.go:117] "RemoveContainer" containerID="c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.612655 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x8lp9"] Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.615595 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x8lp9"] Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.625710 5025 scope.go:117] "RemoveContainer" containerID="431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.643462 5025 scope.go:117] "RemoveContainer" containerID="20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e" Oct 04 10:38:06 crc kubenswrapper[5025]: E1004 10:38:06.644096 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e\": container with ID starting with 20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e not found: ID does not exist" containerID="20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.644137 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e"} err="failed to get container status \"20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e\": rpc error: code = NotFound desc = could not find container \"20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e\": container with ID starting with 20dbd7a7c0c6409f9ba034f4cdda07cb8b115fb7a7187392f661cc0c211be60e not found: ID does not exist" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.644663 5025 scope.go:117] "RemoveContainer" containerID="c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd" Oct 04 10:38:06 crc kubenswrapper[5025]: E1004 10:38:06.645475 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd\": container with ID starting with c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd not found: ID does not exist" containerID="c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.645517 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd"} err="failed to get container status \"c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd\": rpc error: code = NotFound desc = could not find container \"c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd\": container with ID starting with c996f77bf650b5348ed9a37dd00dfb3e717c83af6d1e45fdce30f4ad7d60f6dd not found: ID does not exist" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.645550 5025 scope.go:117] "RemoveContainer" containerID="431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582" Oct 04 10:38:06 crc kubenswrapper[5025]: E1004 10:38:06.645971 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582\": container with ID starting with 431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582 not found: ID does not exist" containerID="431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582" Oct 04 10:38:06 crc kubenswrapper[5025]: I1004 10:38:06.646075 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582"} err="failed to get container status \"431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582\": rpc error: code = NotFound desc = could not find container \"431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582\": container with ID starting with 431ac1de93bda85233bc3bf98927ec4dcb489f7174ca7909091a4535589d3582 not found: ID does not exist" Oct 04 10:38:07 crc kubenswrapper[5025]: I1004 10:38:07.258441 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-74s25"] Oct 04 10:38:07 crc kubenswrapper[5025]: I1004 10:38:07.258687 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-74s25" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerName="registry-server" containerID="cri-o://843b0ec54670dd31f933f79691c32e2688b7f58d0d98a59f2a22220c8bb3b3b2" gracePeriod=2 Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.418590 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" path="/var/lib/kubelet/pods/37c2c485-27f8-40bf-b96f-20c99e41c354/volumes" Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.607905 5025 generic.go:334] "Generic (PLEG): container finished" podID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerID="843b0ec54670dd31f933f79691c32e2688b7f58d0d98a59f2a22220c8bb3b3b2" exitCode=0 Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.607943 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74s25" event={"ID":"ec5571cf-ee2f-45e9-993b-85ec83a5e42f","Type":"ContainerDied","Data":"843b0ec54670dd31f933f79691c32e2688b7f58d0d98a59f2a22220c8bb3b3b2"} Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.741596 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.820574 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-utilities\") pod \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.820634 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6tgm\" (UniqueName: \"kubernetes.io/projected/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-kube-api-access-t6tgm\") pod \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.820666 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-catalog-content\") pod \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\" (UID: \"ec5571cf-ee2f-45e9-993b-85ec83a5e42f\") " Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.821480 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-utilities" (OuterVolumeSpecName: "utilities") pod "ec5571cf-ee2f-45e9-993b-85ec83a5e42f" (UID: "ec5571cf-ee2f-45e9-993b-85ec83a5e42f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.826784 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-kube-api-access-t6tgm" (OuterVolumeSpecName: "kube-api-access-t6tgm") pod "ec5571cf-ee2f-45e9-993b-85ec83a5e42f" (UID: "ec5571cf-ee2f-45e9-993b-85ec83a5e42f"). InnerVolumeSpecName "kube-api-access-t6tgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.922286 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:08 crc kubenswrapper[5025]: I1004 10:38:08.922335 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6tgm\" (UniqueName: \"kubernetes.io/projected/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-kube-api-access-t6tgm\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:09 crc kubenswrapper[5025]: I1004 10:38:09.226059 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec5571cf-ee2f-45e9-993b-85ec83a5e42f" (UID: "ec5571cf-ee2f-45e9-993b-85ec83a5e42f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:38:09 crc kubenswrapper[5025]: I1004 10:38:09.327139 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec5571cf-ee2f-45e9-993b-85ec83a5e42f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:09 crc kubenswrapper[5025]: I1004 10:38:09.614372 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-74s25" event={"ID":"ec5571cf-ee2f-45e9-993b-85ec83a5e42f","Type":"ContainerDied","Data":"000732ba3a400297bab93c8c84b01523fb8ad98a23e1b100da80476b15cc5fba"} Oct 04 10:38:09 crc kubenswrapper[5025]: I1004 10:38:09.614426 5025 scope.go:117] "RemoveContainer" containerID="843b0ec54670dd31f933f79691c32e2688b7f58d0d98a59f2a22220c8bb3b3b2" Oct 04 10:38:09 crc kubenswrapper[5025]: I1004 10:38:09.614461 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-74s25" Oct 04 10:38:09 crc kubenswrapper[5025]: I1004 10:38:09.633981 5025 scope.go:117] "RemoveContainer" containerID="f35f80e2c521d82a299254e2827250fb7fc9d36dcf99cbc52644263b1f059190" Oct 04 10:38:09 crc kubenswrapper[5025]: I1004 10:38:09.639544 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-74s25"] Oct 04 10:38:09 crc kubenswrapper[5025]: I1004 10:38:09.644050 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-74s25"] Oct 04 10:38:09 crc kubenswrapper[5025]: I1004 10:38:09.666392 5025 scope.go:117] "RemoveContainer" containerID="c871d6001409de5bddd83b8b4eb89083292e6fc16e0ea5e1832c0f806fad729b" Oct 04 10:38:10 crc kubenswrapper[5025]: I1004 10:38:10.417534 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" path="/var/lib/kubelet/pods/ec5571cf-ee2f-45e9-993b-85ec83a5e42f/volumes" Oct 04 10:38:10 crc kubenswrapper[5025]: I1004 10:38:10.769649 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:38:11 crc kubenswrapper[5025]: I1004 10:38:11.138271 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:38:11 crc kubenswrapper[5025]: I1004 10:38:11.312758 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:38:12 crc kubenswrapper[5025]: I1004 10:38:12.480850 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzljn"] Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.054524 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9l4lc"] Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.054819 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9l4lc" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerName="registry-server" containerID="cri-o://7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa" gracePeriod=2 Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.410617 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.591615 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-catalog-content\") pod \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.591695 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2q4b\" (UniqueName: \"kubernetes.io/projected/a28fcea7-e325-4583-a1bb-64f7d80a00f7-kube-api-access-n2q4b\") pod \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.591726 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-utilities\") pod \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\" (UID: \"a28fcea7-e325-4583-a1bb-64f7d80a00f7\") " Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.593680 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-utilities" (OuterVolumeSpecName: "utilities") pod "a28fcea7-e325-4583-a1bb-64f7d80a00f7" (UID: "a28fcea7-e325-4583-a1bb-64f7d80a00f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.604192 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a28fcea7-e325-4583-a1bb-64f7d80a00f7-kube-api-access-n2q4b" (OuterVolumeSpecName: "kube-api-access-n2q4b") pod "a28fcea7-e325-4583-a1bb-64f7d80a00f7" (UID: "a28fcea7-e325-4583-a1bb-64f7d80a00f7"). InnerVolumeSpecName "kube-api-access-n2q4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.641120 5025 generic.go:334] "Generic (PLEG): container finished" podID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerID="7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa" exitCode=0 Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.641163 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9l4lc" event={"ID":"a28fcea7-e325-4583-a1bb-64f7d80a00f7","Type":"ContainerDied","Data":"7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa"} Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.641179 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9l4lc" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.641204 5025 scope.go:117] "RemoveContainer" containerID="7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.641191 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9l4lc" event={"ID":"a28fcea7-e325-4583-a1bb-64f7d80a00f7","Type":"ContainerDied","Data":"e18e15a00de83ed8e6a7884393d8576520ce2deaafd01c5777ac98d9631e6e73"} Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.652768 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a28fcea7-e325-4583-a1bb-64f7d80a00f7" (UID: "a28fcea7-e325-4583-a1bb-64f7d80a00f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.669591 5025 scope.go:117] "RemoveContainer" containerID="162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.692802 5025 scope.go:117] "RemoveContainer" containerID="0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.693396 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.693417 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2q4b\" (UniqueName: \"kubernetes.io/projected/a28fcea7-e325-4583-a1bb-64f7d80a00f7-kube-api-access-n2q4b\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.693426 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a28fcea7-e325-4583-a1bb-64f7d80a00f7-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.711914 5025 scope.go:117] "RemoveContainer" containerID="7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa" Oct 04 10:38:13 crc kubenswrapper[5025]: E1004 10:38:13.712357 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa\": container with ID starting with 7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa not found: ID does not exist" containerID="7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.712393 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa"} err="failed to get container status \"7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa\": rpc error: code = NotFound desc = could not find container \"7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa\": container with ID starting with 7b2abd0bfc82dd2f34b0830b34e3521e8feab9a562551326f51a95c608608dfa not found: ID does not exist" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.712414 5025 scope.go:117] "RemoveContainer" containerID="162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4" Oct 04 10:38:13 crc kubenswrapper[5025]: E1004 10:38:13.712686 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4\": container with ID starting with 162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4 not found: ID does not exist" containerID="162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.712725 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4"} err="failed to get container status \"162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4\": rpc error: code = NotFound desc = could not find container \"162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4\": container with ID starting with 162b9d346b03f257b566fdcaab331fbaa8fadeba6a2e7c83ba0757dae39486c4 not found: ID does not exist" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.712750 5025 scope.go:117] "RemoveContainer" containerID="0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f" Oct 04 10:38:13 crc kubenswrapper[5025]: E1004 10:38:13.712982 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f\": container with ID starting with 0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f not found: ID does not exist" containerID="0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.713026 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f"} err="failed to get container status \"0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f\": rpc error: code = NotFound desc = could not find container \"0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f\": container with ID starting with 0162f69f3253e7875c1b4f352dbd249b1caa06638aabbca7c501f12a9a69e54f not found: ID does not exist" Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.970544 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9l4lc"] Oct 04 10:38:13 crc kubenswrapper[5025]: I1004 10:38:13.973163 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9l4lc"] Oct 04 10:38:14 crc kubenswrapper[5025]: I1004 10:38:14.417554 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" path="/var/lib/kubelet/pods/a28fcea7-e325-4583-a1bb-64f7d80a00f7/volumes" Oct 04 10:38:14 crc kubenswrapper[5025]: I1004 10:38:14.713248 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:38:14 crc kubenswrapper[5025]: I1004 10:38:14.713299 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:38:14 crc kubenswrapper[5025]: I1004 10:38:14.713333 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:38:14 crc kubenswrapper[5025]: I1004 10:38:14.713701 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:38:14 crc kubenswrapper[5025]: I1004 10:38:14.713758 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b" gracePeriod=600 Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.258833 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n2pvj"] Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.259491 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n2pvj" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerName="registry-server" containerID="cri-o://7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877" gracePeriod=2 Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.619264 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.655787 5025 generic.go:334] "Generic (PLEG): container finished" podID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerID="7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877" exitCode=0 Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.656084 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pvj" event={"ID":"ef617b3d-a08d-4482-abd5-06a66cb36f45","Type":"ContainerDied","Data":"7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877"} Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.656177 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n2pvj" event={"ID":"ef617b3d-a08d-4482-abd5-06a66cb36f45","Type":"ContainerDied","Data":"093fdc339f8e39167f753e6da6556f00138e047b221818a28e309e5be8912710"} Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.656253 5025 scope.go:117] "RemoveContainer" containerID="7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.656453 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n2pvj" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.659857 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b" exitCode=0 Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.660082 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b"} Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.660169 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"db2ae36648521dea01dfdc74730ff2492700a6b36503b7ffdda74b5fe8a3c056"} Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.684680 5025 scope.go:117] "RemoveContainer" containerID="3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.715615 5025 scope.go:117] "RemoveContainer" containerID="9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.719474 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-utilities\") pod \"ef617b3d-a08d-4482-abd5-06a66cb36f45\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.719595 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t6z5\" (UniqueName: \"kubernetes.io/projected/ef617b3d-a08d-4482-abd5-06a66cb36f45-kube-api-access-5t6z5\") pod \"ef617b3d-a08d-4482-abd5-06a66cb36f45\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.719617 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-catalog-content\") pod \"ef617b3d-a08d-4482-abd5-06a66cb36f45\" (UID: \"ef617b3d-a08d-4482-abd5-06a66cb36f45\") " Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.720484 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-utilities" (OuterVolumeSpecName: "utilities") pod "ef617b3d-a08d-4482-abd5-06a66cb36f45" (UID: "ef617b3d-a08d-4482-abd5-06a66cb36f45"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.728898 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef617b3d-a08d-4482-abd5-06a66cb36f45-kube-api-access-5t6z5" (OuterVolumeSpecName: "kube-api-access-5t6z5") pod "ef617b3d-a08d-4482-abd5-06a66cb36f45" (UID: "ef617b3d-a08d-4482-abd5-06a66cb36f45"). InnerVolumeSpecName "kube-api-access-5t6z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.735596 5025 scope.go:117] "RemoveContainer" containerID="7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877" Oct 04 10:38:15 crc kubenswrapper[5025]: E1004 10:38:15.735991 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877\": container with ID starting with 7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877 not found: ID does not exist" containerID="7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.736030 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877"} err="failed to get container status \"7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877\": rpc error: code = NotFound desc = could not find container \"7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877\": container with ID starting with 7b0d36a5bed19d1ffce7945cb8a2e5404d58e796219ea7fe387107efd7860877 not found: ID does not exist" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.736050 5025 scope.go:117] "RemoveContainer" containerID="3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874" Oct 04 10:38:15 crc kubenswrapper[5025]: E1004 10:38:15.736312 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874\": container with ID starting with 3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874 not found: ID does not exist" containerID="3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.736333 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874"} err="failed to get container status \"3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874\": rpc error: code = NotFound desc = could not find container \"3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874\": container with ID starting with 3e07cbedb3d9cedaf6998c6958c219522595d136f854da81edb8b1e37e4f8874 not found: ID does not exist" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.736345 5025 scope.go:117] "RemoveContainer" containerID="9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075" Oct 04 10:38:15 crc kubenswrapper[5025]: E1004 10:38:15.736603 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075\": container with ID starting with 9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075 not found: ID does not exist" containerID="9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.736625 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075"} err="failed to get container status \"9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075\": rpc error: code = NotFound desc = could not find container \"9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075\": container with ID starting with 9cca6f1610efd95aea838aaecece7347321394f3eaaa53a3ed91ca6f22175075 not found: ID does not exist" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.773356 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef617b3d-a08d-4482-abd5-06a66cb36f45" (UID: "ef617b3d-a08d-4482-abd5-06a66cb36f45"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.821494 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.821772 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef617b3d-a08d-4482-abd5-06a66cb36f45-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.821782 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t6z5\" (UniqueName: \"kubernetes.io/projected/ef617b3d-a08d-4482-abd5-06a66cb36f45-kube-api-access-5t6z5\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.983823 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n2pvj"] Oct 04 10:38:15 crc kubenswrapper[5025]: I1004 10:38:15.986663 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n2pvj"] Oct 04 10:38:16 crc kubenswrapper[5025]: I1004 10:38:16.417390 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" path="/var/lib/kubelet/pods/ef617b3d-a08d-4482-abd5-06a66cb36f45/volumes" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.521209 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" podUID="7a2c8998-2de1-453a-9adb-72facdc8bd2f" containerName="oauth-openshift" containerID="cri-o://b10e0caa164986d35464f6fe88fecb7214e3a05ad09a1201b0d82c82b98c8be3" gracePeriod=15 Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.785933 5025 generic.go:334] "Generic (PLEG): container finished" podID="7a2c8998-2de1-453a-9adb-72facdc8bd2f" containerID="b10e0caa164986d35464f6fe88fecb7214e3a05ad09a1201b0d82c82b98c8be3" exitCode=0 Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.786305 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" event={"ID":"7a2c8998-2de1-453a-9adb-72facdc8bd2f","Type":"ContainerDied","Data":"b10e0caa164986d35464f6fe88fecb7214e3a05ad09a1201b0d82c82b98c8be3"} Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.901935 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.933923 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-login\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934007 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-service-ca\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934082 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-ocp-branding-template\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934124 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-session\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934176 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-policies\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934219 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-router-certs\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934273 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-error\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934313 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-idp-0-file-data\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934383 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-trusted-ca-bundle\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934416 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-dir\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934452 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-provider-selection\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934491 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-cliconfig\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934533 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g7tl\" (UniqueName: \"kubernetes.io/projected/7a2c8998-2de1-453a-9adb-72facdc8bd2f-kube-api-access-9g7tl\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.934566 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-serving-cert\") pod \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\" (UID: \"7a2c8998-2de1-453a-9adb-72facdc8bd2f\") " Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.940726 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-68974c876c-6fqrc"] Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941087 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerName="extract-content" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941147 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerName="extract-content" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941164 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941172 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941205 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerName="extract-utilities" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941216 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerName="extract-utilities" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941227 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941234 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941245 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941253 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941284 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b83ab4d7-ed5d-4d43-b723-b1c401465fb0" containerName="pruner" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941294 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b83ab4d7-ed5d-4d43-b723-b1c401465fb0" containerName="pruner" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941307 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a2c8998-2de1-453a-9adb-72facdc8bd2f" containerName="oauth-openshift" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941315 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a2c8998-2de1-453a-9adb-72facdc8bd2f" containerName="oauth-openshift" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941328 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerName="extract-content" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941336 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerName="extract-content" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941369 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerName="extract-content" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941380 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerName="extract-content" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941392 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerName="extract-content" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941400 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerName="extract-content" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941410 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerName="extract-utilities" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941419 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerName="extract-utilities" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941457 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941469 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941496 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerName="extract-utilities" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941506 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerName="extract-utilities" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941547 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd" containerName="pruner" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941557 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd" containerName="pruner" Oct 04 10:38:37 crc kubenswrapper[5025]: E1004 10:38:37.941572 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerName="extract-utilities" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941581 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerName="extract-utilities" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941780 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="a28fcea7-e325-4583-a1bb-64f7d80a00f7" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941798 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef617b3d-a08d-4482-abd5-06a66cb36f45" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941807 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a2c8998-2de1-453a-9adb-72facdc8bd2f" containerName="oauth-openshift" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941824 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="b83ab4d7-ed5d-4d43-b723-b1c401465fb0" containerName="pruner" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941835 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec5571cf-ee2f-45e9-993b-85ec83a5e42f" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941870 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef07b9a8-7ae9-4e6e-8189-bde3d75d7ddd" containerName="pruner" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.941884 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="37c2c485-27f8-40bf-b96f-20c99e41c354" containerName="registry-server" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.942489 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.945087 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.945414 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.946167 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.946556 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.948178 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.987205 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.987595 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.987807 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.988041 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.988077 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a2c8998-2de1-453a-9adb-72facdc8bd2f-kube-api-access-9g7tl" (OuterVolumeSpecName: "kube-api-access-9g7tl") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "kube-api-access-9g7tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.988749 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.989347 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.993914 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:38:37 crc kubenswrapper[5025]: I1004 10:38:37.994082 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7a2c8998-2de1-453a-9adb-72facdc8bd2f" (UID: "7a2c8998-2de1-453a-9adb-72facdc8bd2f"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.011665 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68974c876c-6fqrc"] Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.035625 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.035676 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-router-certs\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.035734 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.035758 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s8wk\" (UniqueName: \"kubernetes.io/projected/6441b8ee-99b0-4b63-b546-11aa74d30268-kube-api-access-2s8wk\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.035790 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-service-ca\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.035822 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-session\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.035857 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.035893 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.035920 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-audit-policies\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.035990 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-template-login\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036035 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036060 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-template-error\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036082 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036099 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6441b8ee-99b0-4b63-b546-11aa74d30268-audit-dir\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036154 5025 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036167 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036177 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036186 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036212 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036222 5025 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7a2c8998-2de1-453a-9adb-72facdc8bd2f-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036232 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036241 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036250 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g7tl\" (UniqueName: \"kubernetes.io/projected/7a2c8998-2de1-453a-9adb-72facdc8bd2f-kube-api-access-9g7tl\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036259 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036268 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036277 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036286 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.036295 5025 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7a2c8998-2de1-453a-9adb-72facdc8bd2f-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137214 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137297 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137335 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-audit-policies\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137401 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-template-login\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137431 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137470 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-template-error\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137509 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6441b8ee-99b0-4b63-b546-11aa74d30268-audit-dir\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137540 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137594 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137635 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-router-certs\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137641 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6441b8ee-99b0-4b63-b546-11aa74d30268-audit-dir\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137669 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137701 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s8wk\" (UniqueName: \"kubernetes.io/projected/6441b8ee-99b0-4b63-b546-11aa74d30268-kube-api-access-2s8wk\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137734 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-service-ca\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.137780 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-session\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.138234 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-audit-policies\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.138265 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-cliconfig\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.139122 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-service-ca\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.139196 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.141548 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-template-login\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.141781 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.141841 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-serving-cert\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.142112 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.142686 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-user-template-error\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.143950 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-session\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.144114 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.144336 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6441b8ee-99b0-4b63-b546-11aa74d30268-v4-0-config-system-router-certs\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.155080 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s8wk\" (UniqueName: \"kubernetes.io/projected/6441b8ee-99b0-4b63-b546-11aa74d30268-kube-api-access-2s8wk\") pod \"oauth-openshift-68974c876c-6fqrc\" (UID: \"6441b8ee-99b0-4b63-b546-11aa74d30268\") " pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.325695 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.795867 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" event={"ID":"7a2c8998-2de1-453a-9adb-72facdc8bd2f","Type":"ContainerDied","Data":"9766ce66c8de239c1c107e3f70622944739f26b464ca39b9b8da45007552fffe"} Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.796389 5025 scope.go:117] "RemoveContainer" containerID="b10e0caa164986d35464f6fe88fecb7214e3a05ad09a1201b0d82c82b98c8be3" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.796614 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bzljn" Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.799741 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-68974c876c-6fqrc"] Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.831718 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzljn"] Oct 04 10:38:38 crc kubenswrapper[5025]: I1004 10:38:38.837644 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bzljn"] Oct 04 10:38:39 crc kubenswrapper[5025]: I1004 10:38:39.804969 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" event={"ID":"6441b8ee-99b0-4b63-b546-11aa74d30268","Type":"ContainerStarted","Data":"678a37ac23413ba5d72a03d5bc6fa3cb8196ca03ef7ac4669bc96b8ac9647bc2"} Oct 04 10:38:39 crc kubenswrapper[5025]: I1004 10:38:39.805274 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" event={"ID":"6441b8ee-99b0-4b63-b546-11aa74d30268","Type":"ContainerStarted","Data":"766ba21a953f33148e4f91729f252c30307b3df18969e6d759291ac8191a4667"} Oct 04 10:38:39 crc kubenswrapper[5025]: I1004 10:38:39.806303 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:39 crc kubenswrapper[5025]: I1004 10:38:39.813703 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" Oct 04 10:38:39 crc kubenswrapper[5025]: I1004 10:38:39.830815 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-68974c876c-6fqrc" podStartSLOduration=27.830787936 podStartE2EDuration="27.830787936s" podCreationTimestamp="2025-10-04 10:38:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:38:39.82939165 +0000 UTC m=+248.254358580" watchObservedRunningTime="2025-10-04 10:38:39.830787936 +0000 UTC m=+248.255754856" Oct 04 10:38:40 crc kubenswrapper[5025]: I1004 10:38:40.419210 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a2c8998-2de1-453a-9adb-72facdc8bd2f" path="/var/lib/kubelet/pods/7a2c8998-2de1-453a-9adb-72facdc8bd2f/volumes" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.308811 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8k7kn"] Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.309843 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8k7kn" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerName="registry-server" containerID="cri-o://25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f" gracePeriod=30 Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.339570 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6xkcs"] Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.339963 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6xkcs" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerName="registry-server" containerID="cri-o://5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809" gracePeriod=30 Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.347497 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sp4mz"] Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.348360 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" podUID="24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" containerName="marketplace-operator" containerID="cri-o://31be2214ed9ede5b72baef4d475d25d0c524b556cf3b5e322ddd4f38492570fb" gracePeriod=30 Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.355618 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2fdt"] Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.355897 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-p2fdt" podUID="35966afc-f931-4779-a1bc-88b64819f756" containerName="registry-server" containerID="cri-o://6fbe8530238f916aa361354131440486dbe0c47ca8c4fdb3a7e5a127017cc8b5" gracePeriod=30 Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.374772 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwws6"] Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.375099 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xwws6" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerName="registry-server" containerID="cri-o://8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340" gracePeriod=30 Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.376059 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d2str"] Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.377245 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.388689 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d2str"] Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.477371 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/894ba87a-0076-47c2-836d-c972664c8900-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d2str\" (UID: \"894ba87a-0076-47c2-836d-c972664c8900\") " pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.477701 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qr9w\" (UniqueName: \"kubernetes.io/projected/894ba87a-0076-47c2-836d-c972664c8900-kube-api-access-4qr9w\") pod \"marketplace-operator-79b997595-d2str\" (UID: \"894ba87a-0076-47c2-836d-c972664c8900\") " pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.477742 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/894ba87a-0076-47c2-836d-c972664c8900-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d2str\" (UID: \"894ba87a-0076-47c2-836d-c972664c8900\") " pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.578303 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/894ba87a-0076-47c2-836d-c972664c8900-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d2str\" (UID: \"894ba87a-0076-47c2-836d-c972664c8900\") " pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.578367 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qr9w\" (UniqueName: \"kubernetes.io/projected/894ba87a-0076-47c2-836d-c972664c8900-kube-api-access-4qr9w\") pod \"marketplace-operator-79b997595-d2str\" (UID: \"894ba87a-0076-47c2-836d-c972664c8900\") " pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.578397 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/894ba87a-0076-47c2-836d-c972664c8900-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d2str\" (UID: \"894ba87a-0076-47c2-836d-c972664c8900\") " pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.579857 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/894ba87a-0076-47c2-836d-c972664c8900-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d2str\" (UID: \"894ba87a-0076-47c2-836d-c972664c8900\") " pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.585101 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/894ba87a-0076-47c2-836d-c972664c8900-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d2str\" (UID: \"894ba87a-0076-47c2-836d-c972664c8900\") " pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.595380 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qr9w\" (UniqueName: \"kubernetes.io/projected/894ba87a-0076-47c2-836d-c972664c8900-kube-api-access-4qr9w\") pod \"marketplace-operator-79b997595-d2str\" (UID: \"894ba87a-0076-47c2-836d-c972664c8900\") " pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.699516 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:40 crc kubenswrapper[5025]: E1004 10:39:40.715424 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f is running failed: container process not found" containerID="25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 10:39:40 crc kubenswrapper[5025]: E1004 10:39:40.715955 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f is running failed: container process not found" containerID="25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 10:39:40 crc kubenswrapper[5025]: E1004 10:39:40.716497 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f is running failed: container process not found" containerID="25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 10:39:40 crc kubenswrapper[5025]: E1004 10:39:40.716533 5025 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-8k7kn" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerName="registry-server" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.840769 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.882454 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwsfv\" (UniqueName: \"kubernetes.io/projected/84e2a31a-4153-4a4c-b56b-8480c8662e07-kube-api-access-nwsfv\") pod \"84e2a31a-4153-4a4c-b56b-8480c8662e07\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.882523 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-utilities\") pod \"84e2a31a-4153-4a4c-b56b-8480c8662e07\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.882616 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-catalog-content\") pod \"84e2a31a-4153-4a4c-b56b-8480c8662e07\" (UID: \"84e2a31a-4153-4a4c-b56b-8480c8662e07\") " Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.885949 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-utilities" (OuterVolumeSpecName: "utilities") pod "84e2a31a-4153-4a4c-b56b-8480c8662e07" (UID: "84e2a31a-4153-4a4c-b56b-8480c8662e07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.891776 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84e2a31a-4153-4a4c-b56b-8480c8662e07-kube-api-access-nwsfv" (OuterVolumeSpecName: "kube-api-access-nwsfv") pod "84e2a31a-4153-4a4c-b56b-8480c8662e07" (UID: "84e2a31a-4153-4a4c-b56b-8480c8662e07"). InnerVolumeSpecName "kube-api-access-nwsfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:39:40 crc kubenswrapper[5025]: E1004 10:39:40.903112 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809 is running failed: container process not found" containerID="5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 10:39:40 crc kubenswrapper[5025]: E1004 10:39:40.904245 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809 is running failed: container process not found" containerID="5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 10:39:40 crc kubenswrapper[5025]: E1004 10:39:40.904831 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809 is running failed: container process not found" containerID="5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809" cmd=["grpc_health_probe","-addr=:50051"] Oct 04 10:39:40 crc kubenswrapper[5025]: E1004 10:39:40.904918 5025 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-6xkcs" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerName="registry-server" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.914255 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d2str"] Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.983714 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwsfv\" (UniqueName: \"kubernetes.io/projected/84e2a31a-4153-4a4c-b56b-8480c8662e07-kube-api-access-nwsfv\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:40 crc kubenswrapper[5025]: I1004 10:39:40.983755 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.032206 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84e2a31a-4153-4a4c-b56b-8480c8662e07" (UID: "84e2a31a-4153-4a4c-b56b-8480c8662e07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.084648 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84e2a31a-4153-4a4c-b56b-8480c8662e07-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.224943 5025 generic.go:334] "Generic (PLEG): container finished" podID="24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" containerID="31be2214ed9ede5b72baef4d475d25d0c524b556cf3b5e322ddd4f38492570fb" exitCode=0 Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.225045 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" event={"ID":"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef","Type":"ContainerDied","Data":"31be2214ed9ede5b72baef4d475d25d0c524b556cf3b5e322ddd4f38492570fb"} Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.225084 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" event={"ID":"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef","Type":"ContainerDied","Data":"f3b69e8ca74f9ec00e698996bef1d43fb052a92e8b820475b07e6340f5f40bcc"} Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.225097 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3b69e8ca74f9ec00e698996bef1d43fb052a92e8b820475b07e6340f5f40bcc" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.226797 5025 generic.go:334] "Generic (PLEG): container finished" podID="35966afc-f931-4779-a1bc-88b64819f756" containerID="6fbe8530238f916aa361354131440486dbe0c47ca8c4fdb3a7e5a127017cc8b5" exitCode=0 Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.226886 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2fdt" event={"ID":"35966afc-f931-4779-a1bc-88b64819f756","Type":"ContainerDied","Data":"6fbe8530238f916aa361354131440486dbe0c47ca8c4fdb3a7e5a127017cc8b5"} Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.227576 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.229413 5025 generic.go:334] "Generic (PLEG): container finished" podID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerID="25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f" exitCode=0 Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.229490 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k7kn" event={"ID":"a36bada3-22b5-48f9-b82e-26a30d24f57f","Type":"ContainerDied","Data":"25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f"} Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.230421 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d2str" event={"ID":"894ba87a-0076-47c2-836d-c972664c8900","Type":"ContainerStarted","Data":"30e8c318582cf713e4719e7c541b57191932a85d6c8f3300d4801b23744e2a3c"} Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.232235 5025 generic.go:334] "Generic (PLEG): container finished" podID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerID="5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809" exitCode=0 Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.232322 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xkcs" event={"ID":"ed7e8716-6f98-43ad-aa7a-5a2b0b354091","Type":"ContainerDied","Data":"5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809"} Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.233649 5025 generic.go:334] "Generic (PLEG): container finished" podID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerID="8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340" exitCode=0 Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.233684 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwws6" event={"ID":"84e2a31a-4153-4a4c-b56b-8480c8662e07","Type":"ContainerDied","Data":"8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340"} Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.233706 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwws6" event={"ID":"84e2a31a-4153-4a4c-b56b-8480c8662e07","Type":"ContainerDied","Data":"6734cfda8d5f5d08f103501ae9fdda6f3074a159a868fea6366ea8fd304f7f10"} Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.233718 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwws6" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.233725 5025 scope.go:117] "RemoveContainer" containerID="8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.273874 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwws6"] Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.284722 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.285539 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xwws6"] Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.286960 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2jwv\" (UniqueName: \"kubernetes.io/projected/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-kube-api-access-f2jwv\") pod \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.287533 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-operator-metrics\") pod \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.287609 5025 scope.go:117] "RemoveContainer" containerID="fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.287623 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-trusted-ca\") pod \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\" (UID: \"24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.288135 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" (UID: "24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.292642 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" (UID: "24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.297871 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-kube-api-access-f2jwv" (OuterVolumeSpecName: "kube-api-access-f2jwv") pod "24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" (UID: "24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef"). InnerVolumeSpecName "kube-api-access-f2jwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.308473 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.313472 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.333686 5025 scope.go:117] "RemoveContainer" containerID="ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.360332 5025 scope.go:117] "RemoveContainer" containerID="8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340" Oct 04 10:39:41 crc kubenswrapper[5025]: E1004 10:39:41.361145 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340\": container with ID starting with 8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340 not found: ID does not exist" containerID="8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.361175 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340"} err="failed to get container status \"8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340\": rpc error: code = NotFound desc = could not find container \"8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340\": container with ID starting with 8dae6196f52eaa3dbbe6fe008fecd12ba81f943ab0c5a8d6a405a46d9bbd2340 not found: ID does not exist" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.361195 5025 scope.go:117] "RemoveContainer" containerID="fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392" Oct 04 10:39:41 crc kubenswrapper[5025]: E1004 10:39:41.361826 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392\": container with ID starting with fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392 not found: ID does not exist" containerID="fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.361854 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392"} err="failed to get container status \"fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392\": rpc error: code = NotFound desc = could not find container \"fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392\": container with ID starting with fea427e3984601e1259e74a39ed646a732cc100eff6da7986667d4dbb8276392 not found: ID does not exist" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.361873 5025 scope.go:117] "RemoveContainer" containerID="ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736" Oct 04 10:39:41 crc kubenswrapper[5025]: E1004 10:39:41.362338 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736\": container with ID starting with ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736 not found: ID does not exist" containerID="ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.362355 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736"} err="failed to get container status \"ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736\": rpc error: code = NotFound desc = could not find container \"ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736\": container with ID starting with ded119c70e6aa96be561fc3ffc49b944096ec99237f89c11801e04902d907736 not found: ID does not exist" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.388504 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-utilities\") pod \"a36bada3-22b5-48f9-b82e-26a30d24f57f\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.388561 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdmnp\" (UniqueName: \"kubernetes.io/projected/a36bada3-22b5-48f9-b82e-26a30d24f57f-kube-api-access-bdmnp\") pod \"a36bada3-22b5-48f9-b82e-26a30d24f57f\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.388615 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-catalog-content\") pod \"a36bada3-22b5-48f9-b82e-26a30d24f57f\" (UID: \"a36bada3-22b5-48f9-b82e-26a30d24f57f\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.388688 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-utilities\") pod \"35966afc-f931-4779-a1bc-88b64819f756\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.388712 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-catalog-content\") pod \"35966afc-f931-4779-a1bc-88b64819f756\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.388747 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-catalog-content\") pod \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.388794 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7qhg\" (UniqueName: \"kubernetes.io/projected/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-kube-api-access-c7qhg\") pod \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.388819 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvdv2\" (UniqueName: \"kubernetes.io/projected/35966afc-f931-4779-a1bc-88b64819f756-kube-api-access-qvdv2\") pod \"35966afc-f931-4779-a1bc-88b64819f756\" (UID: \"35966afc-f931-4779-a1bc-88b64819f756\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.388841 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-utilities\") pod \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\" (UID: \"ed7e8716-6f98-43ad-aa7a-5a2b0b354091\") " Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.389113 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2jwv\" (UniqueName: \"kubernetes.io/projected/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-kube-api-access-f2jwv\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.389132 5025 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.389145 5025 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.389791 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-utilities" (OuterVolumeSpecName: "utilities") pod "ed7e8716-6f98-43ad-aa7a-5a2b0b354091" (UID: "ed7e8716-6f98-43ad-aa7a-5a2b0b354091"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.390492 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-utilities" (OuterVolumeSpecName: "utilities") pod "a36bada3-22b5-48f9-b82e-26a30d24f57f" (UID: "a36bada3-22b5-48f9-b82e-26a30d24f57f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.392837 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-utilities" (OuterVolumeSpecName: "utilities") pod "35966afc-f931-4779-a1bc-88b64819f756" (UID: "35966afc-f931-4779-a1bc-88b64819f756"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.393912 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36bada3-22b5-48f9-b82e-26a30d24f57f-kube-api-access-bdmnp" (OuterVolumeSpecName: "kube-api-access-bdmnp") pod "a36bada3-22b5-48f9-b82e-26a30d24f57f" (UID: "a36bada3-22b5-48f9-b82e-26a30d24f57f"). InnerVolumeSpecName "kube-api-access-bdmnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.394974 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-kube-api-access-c7qhg" (OuterVolumeSpecName: "kube-api-access-c7qhg") pod "ed7e8716-6f98-43ad-aa7a-5a2b0b354091" (UID: "ed7e8716-6f98-43ad-aa7a-5a2b0b354091"). InnerVolumeSpecName "kube-api-access-c7qhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.399573 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35966afc-f931-4779-a1bc-88b64819f756-kube-api-access-qvdv2" (OuterVolumeSpecName: "kube-api-access-qvdv2") pod "35966afc-f931-4779-a1bc-88b64819f756" (UID: "35966afc-f931-4779-a1bc-88b64819f756"). InnerVolumeSpecName "kube-api-access-qvdv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.405791 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35966afc-f931-4779-a1bc-88b64819f756" (UID: "35966afc-f931-4779-a1bc-88b64819f756"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.455641 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed7e8716-6f98-43ad-aa7a-5a2b0b354091" (UID: "ed7e8716-6f98-43ad-aa7a-5a2b0b354091"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.463971 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a36bada3-22b5-48f9-b82e-26a30d24f57f" (UID: "a36bada3-22b5-48f9-b82e-26a30d24f57f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.490074 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.490110 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35966afc-f931-4779-a1bc-88b64819f756-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.490128 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.490140 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7qhg\" (UniqueName: \"kubernetes.io/projected/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-kube-api-access-c7qhg\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.490154 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvdv2\" (UniqueName: \"kubernetes.io/projected/35966afc-f931-4779-a1bc-88b64819f756-kube-api-access-qvdv2\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.490165 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed7e8716-6f98-43ad-aa7a-5a2b0b354091-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.490178 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.490189 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdmnp\" (UniqueName: \"kubernetes.io/projected/a36bada3-22b5-48f9-b82e-26a30d24f57f-kube-api-access-bdmnp\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:41 crc kubenswrapper[5025]: I1004 10:39:41.490201 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a36bada3-22b5-48f9-b82e-26a30d24f57f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.244843 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d2str" event={"ID":"894ba87a-0076-47c2-836d-c972664c8900","Type":"ContainerStarted","Data":"767ce4460dc5d1adfcff2b7f8605803edb0785b73265def8961b0ccda971aa10"} Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.245389 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.249126 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6xkcs" event={"ID":"ed7e8716-6f98-43ad-aa7a-5a2b0b354091","Type":"ContainerDied","Data":"d53c12b9d2eb10b19c9b6daa1a49f77d605c5554b9891c20ea37b9504912c7dc"} Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.249189 5025 scope.go:117] "RemoveContainer" containerID="5bec20f3cdad9788642d9fac79d00d281176a70e8b772a2011b57b003928a809" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.249317 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6xkcs" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.253440 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-d2str" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.259212 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2fdt" event={"ID":"35966afc-f931-4779-a1bc-88b64819f756","Type":"ContainerDied","Data":"f56a9eacc2fe5811e570d27a04979fadfa9786ff4e0b2be38d5b2aa2f9059a00"} Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.259310 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2fdt" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.275037 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sp4mz" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.276538 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8k7kn" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.278827 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8k7kn" event={"ID":"a36bada3-22b5-48f9-b82e-26a30d24f57f","Type":"ContainerDied","Data":"2c6ded6cdd40b108760e5c77e027bce33aa708ddf83c0ec119751a8d02eec96d"} Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.285531 5025 scope.go:117] "RemoveContainer" containerID="5ed121002a57ccad828494e90f66381f572564232e88cf63ccdc074b425fc83b" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.290102 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-d2str" podStartSLOduration=2.290084779 podStartE2EDuration="2.290084779s" podCreationTimestamp="2025-10-04 10:39:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:39:42.273778555 +0000 UTC m=+310.698745445" watchObservedRunningTime="2025-10-04 10:39:42.290084779 +0000 UTC m=+310.715051659" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.293622 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6xkcs"] Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.302037 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6xkcs"] Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.322107 5025 scope.go:117] "RemoveContainer" containerID="2391aab9f62481da328b20032accfb573c00df02062f8db7f650483b7d3d28c5" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.347569 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2fdt"] Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.350099 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2fdt"] Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.369998 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sp4mz"] Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.375664 5025 scope.go:117] "RemoveContainer" containerID="6fbe8530238f916aa361354131440486dbe0c47ca8c4fdb3a7e5a127017cc8b5" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.375863 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sp4mz"] Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.383074 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8k7kn"] Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.387399 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8k7kn"] Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.397738 5025 scope.go:117] "RemoveContainer" containerID="a4d663f3225ebb067abd8feb86595c927c085e2192841892dfaf195a57992270" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.413133 5025 scope.go:117] "RemoveContainer" containerID="3e8d7330cd9de49308d8f2da450b86f30fe2a44a32f9554d9899ddfd5b2440fc" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.421287 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" path="/var/lib/kubelet/pods/24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef/volumes" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.421897 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35966afc-f931-4779-a1bc-88b64819f756" path="/var/lib/kubelet/pods/35966afc-f931-4779-a1bc-88b64819f756/volumes" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.422788 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" path="/var/lib/kubelet/pods/84e2a31a-4153-4a4c-b56b-8480c8662e07/volumes" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.424222 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" path="/var/lib/kubelet/pods/a36bada3-22b5-48f9-b82e-26a30d24f57f/volumes" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.424884 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" path="/var/lib/kubelet/pods/ed7e8716-6f98-43ad-aa7a-5a2b0b354091/volumes" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.432259 5025 scope.go:117] "RemoveContainer" containerID="25f310569521097c3483e1de2ffa1711a4ce6667a7769369d5827e9f5ebf769f" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.445301 5025 scope.go:117] "RemoveContainer" containerID="01bfa63851deb6258320471d7ff712e86219ddff178d622f01280c0ec060acf0" Oct 04 10:39:42 crc kubenswrapper[5025]: I1004 10:39:42.460753 5025 scope.go:117] "RemoveContainer" containerID="994bb912dd26aab48054dea9ca713672054ba76aaca69658676671c1d58a7f98" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.135005 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m9hl2"] Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.135757 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerName="extract-content" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.135790 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerName="extract-content" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.135813 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" containerName="marketplace-operator" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.135862 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" containerName="marketplace-operator" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.135884 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.135897 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.135917 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerName="extract-utilities" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.135929 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerName="extract-utilities" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.135952 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerName="extract-utilities" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.135964 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerName="extract-utilities" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.135999 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerName="extract-content" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136038 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerName="extract-content" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.136056 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136068 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.136087 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136101 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.136119 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35966afc-f931-4779-a1bc-88b64819f756" containerName="extract-utilities" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136131 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="35966afc-f931-4779-a1bc-88b64819f756" containerName="extract-utilities" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.136147 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerName="extract-content" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136159 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerName="extract-content" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.136179 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35966afc-f931-4779-a1bc-88b64819f756" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136192 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="35966afc-f931-4779-a1bc-88b64819f756" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.136211 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerName="extract-utilities" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136223 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerName="extract-utilities" Oct 04 10:39:43 crc kubenswrapper[5025]: E1004 10:39:43.136242 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35966afc-f931-4779-a1bc-88b64819f756" containerName="extract-content" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136254 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="35966afc-f931-4779-a1bc-88b64819f756" containerName="extract-content" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136431 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="24cdcbf7-e6c5-4e50-b002-8beb6c1a41ef" containerName="marketplace-operator" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136455 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="84e2a31a-4153-4a4c-b56b-8480c8662e07" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136474 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed7e8716-6f98-43ad-aa7a-5a2b0b354091" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136494 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="35966afc-f931-4779-a1bc-88b64819f756" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.136508 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36bada3-22b5-48f9-b82e-26a30d24f57f" containerName="registry-server" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.142676 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.145644 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.145852 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m9hl2"] Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.208156 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e7faef-f9a9-4742-b6e1-893e70ab32b3-catalog-content\") pod \"redhat-marketplace-m9hl2\" (UID: \"c3e7faef-f9a9-4742-b6e1-893e70ab32b3\") " pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.208240 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e7faef-f9a9-4742-b6e1-893e70ab32b3-utilities\") pod \"redhat-marketplace-m9hl2\" (UID: \"c3e7faef-f9a9-4742-b6e1-893e70ab32b3\") " pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.208788 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhhtw\" (UniqueName: \"kubernetes.io/projected/c3e7faef-f9a9-4742-b6e1-893e70ab32b3-kube-api-access-jhhtw\") pod \"redhat-marketplace-m9hl2\" (UID: \"c3e7faef-f9a9-4742-b6e1-893e70ab32b3\") " pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.310473 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhhtw\" (UniqueName: \"kubernetes.io/projected/c3e7faef-f9a9-4742-b6e1-893e70ab32b3-kube-api-access-jhhtw\") pod \"redhat-marketplace-m9hl2\" (UID: \"c3e7faef-f9a9-4742-b6e1-893e70ab32b3\") " pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.310660 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e7faef-f9a9-4742-b6e1-893e70ab32b3-catalog-content\") pod \"redhat-marketplace-m9hl2\" (UID: \"c3e7faef-f9a9-4742-b6e1-893e70ab32b3\") " pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.310724 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e7faef-f9a9-4742-b6e1-893e70ab32b3-utilities\") pod \"redhat-marketplace-m9hl2\" (UID: \"c3e7faef-f9a9-4742-b6e1-893e70ab32b3\") " pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.311486 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e7faef-f9a9-4742-b6e1-893e70ab32b3-utilities\") pod \"redhat-marketplace-m9hl2\" (UID: \"c3e7faef-f9a9-4742-b6e1-893e70ab32b3\") " pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.313736 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e7faef-f9a9-4742-b6e1-893e70ab32b3-catalog-content\") pod \"redhat-marketplace-m9hl2\" (UID: \"c3e7faef-f9a9-4742-b6e1-893e70ab32b3\") " pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.342830 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhhtw\" (UniqueName: \"kubernetes.io/projected/c3e7faef-f9a9-4742-b6e1-893e70ab32b3-kube-api-access-jhhtw\") pod \"redhat-marketplace-m9hl2\" (UID: \"c3e7faef-f9a9-4742-b6e1-893e70ab32b3\") " pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.484312 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:43 crc kubenswrapper[5025]: I1004 10:39:43.729615 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m9hl2"] Oct 04 10:39:43 crc kubenswrapper[5025]: W1004 10:39:43.735731 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3e7faef_f9a9_4742_b6e1_893e70ab32b3.slice/crio-d8eecb7e495782ccdcbebeaa002eb59fd367f0c85c582c8a1a60f06ec718b877 WatchSource:0}: Error finding container d8eecb7e495782ccdcbebeaa002eb59fd367f0c85c582c8a1a60f06ec718b877: Status 404 returned error can't find the container with id d8eecb7e495782ccdcbebeaa002eb59fd367f0c85c582c8a1a60f06ec718b877 Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.133222 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rxpjg"] Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.136497 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.137942 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rxpjg"] Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.138678 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.224022 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b1dd0c3-486f-4ac8-be1b-cfc100de913e-utilities\") pod \"redhat-operators-rxpjg\" (UID: \"1b1dd0c3-486f-4ac8-be1b-cfc100de913e\") " pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.224308 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b1dd0c3-486f-4ac8-be1b-cfc100de913e-catalog-content\") pod \"redhat-operators-rxpjg\" (UID: \"1b1dd0c3-486f-4ac8-be1b-cfc100de913e\") " pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.224390 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-474z6\" (UniqueName: \"kubernetes.io/projected/1b1dd0c3-486f-4ac8-be1b-cfc100de913e-kube-api-access-474z6\") pod \"redhat-operators-rxpjg\" (UID: \"1b1dd0c3-486f-4ac8-be1b-cfc100de913e\") " pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.297155 5025 generic.go:334] "Generic (PLEG): container finished" podID="c3e7faef-f9a9-4742-b6e1-893e70ab32b3" containerID="31ae5ba7501e8fe726c95f6eb1a8dbe1347c6b5b1dfd33cd17a0dfa30c476028" exitCode=0 Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.297250 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9hl2" event={"ID":"c3e7faef-f9a9-4742-b6e1-893e70ab32b3","Type":"ContainerDied","Data":"31ae5ba7501e8fe726c95f6eb1a8dbe1347c6b5b1dfd33cd17a0dfa30c476028"} Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.297382 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9hl2" event={"ID":"c3e7faef-f9a9-4742-b6e1-893e70ab32b3","Type":"ContainerStarted","Data":"d8eecb7e495782ccdcbebeaa002eb59fd367f0c85c582c8a1a60f06ec718b877"} Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.325556 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b1dd0c3-486f-4ac8-be1b-cfc100de913e-catalog-content\") pod \"redhat-operators-rxpjg\" (UID: \"1b1dd0c3-486f-4ac8-be1b-cfc100de913e\") " pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.325648 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-474z6\" (UniqueName: \"kubernetes.io/projected/1b1dd0c3-486f-4ac8-be1b-cfc100de913e-kube-api-access-474z6\") pod \"redhat-operators-rxpjg\" (UID: \"1b1dd0c3-486f-4ac8-be1b-cfc100de913e\") " pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.325688 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b1dd0c3-486f-4ac8-be1b-cfc100de913e-utilities\") pod \"redhat-operators-rxpjg\" (UID: \"1b1dd0c3-486f-4ac8-be1b-cfc100de913e\") " pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.326398 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b1dd0c3-486f-4ac8-be1b-cfc100de913e-utilities\") pod \"redhat-operators-rxpjg\" (UID: \"1b1dd0c3-486f-4ac8-be1b-cfc100de913e\") " pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.326465 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b1dd0c3-486f-4ac8-be1b-cfc100de913e-catalog-content\") pod \"redhat-operators-rxpjg\" (UID: \"1b1dd0c3-486f-4ac8-be1b-cfc100de913e\") " pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.347272 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-474z6\" (UniqueName: \"kubernetes.io/projected/1b1dd0c3-486f-4ac8-be1b-cfc100de913e-kube-api-access-474z6\") pod \"redhat-operators-rxpjg\" (UID: \"1b1dd0c3-486f-4ac8-be1b-cfc100de913e\") " pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.455237 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:44 crc kubenswrapper[5025]: I1004 10:39:44.629566 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rxpjg"] Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.307004 5025 generic.go:334] "Generic (PLEG): container finished" podID="1b1dd0c3-486f-4ac8-be1b-cfc100de913e" containerID="4122d3e5268eba8b90eefef88bf6b8873bf009e9e496a9c6ded98916bd5af93a" exitCode=0 Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.307074 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxpjg" event={"ID":"1b1dd0c3-486f-4ac8-be1b-cfc100de913e","Type":"ContainerDied","Data":"4122d3e5268eba8b90eefef88bf6b8873bf009e9e496a9c6ded98916bd5af93a"} Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.307099 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxpjg" event={"ID":"1b1dd0c3-486f-4ac8-be1b-cfc100de913e","Type":"ContainerStarted","Data":"5f4596ed1f0720b64d11d87c920d6844398a61972187570f62b50c287d42f459"} Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.525139 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d2hs7"] Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.526690 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.530073 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.541314 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d2hs7"] Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.542823 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k5f7\" (UniqueName: \"kubernetes.io/projected/a4a6c356-d827-4b99-a7f2-fbfba013450b-kube-api-access-9k5f7\") pod \"community-operators-d2hs7\" (UID: \"a4a6c356-d827-4b99-a7f2-fbfba013450b\") " pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.543048 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a6c356-d827-4b99-a7f2-fbfba013450b-utilities\") pod \"community-operators-d2hs7\" (UID: \"a4a6c356-d827-4b99-a7f2-fbfba013450b\") " pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.543155 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a6c356-d827-4b99-a7f2-fbfba013450b-catalog-content\") pod \"community-operators-d2hs7\" (UID: \"a4a6c356-d827-4b99-a7f2-fbfba013450b\") " pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.644244 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a6c356-d827-4b99-a7f2-fbfba013450b-catalog-content\") pod \"community-operators-d2hs7\" (UID: \"a4a6c356-d827-4b99-a7f2-fbfba013450b\") " pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.644306 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k5f7\" (UniqueName: \"kubernetes.io/projected/a4a6c356-d827-4b99-a7f2-fbfba013450b-kube-api-access-9k5f7\") pod \"community-operators-d2hs7\" (UID: \"a4a6c356-d827-4b99-a7f2-fbfba013450b\") " pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.644354 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a6c356-d827-4b99-a7f2-fbfba013450b-utilities\") pod \"community-operators-d2hs7\" (UID: \"a4a6c356-d827-4b99-a7f2-fbfba013450b\") " pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.644846 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4a6c356-d827-4b99-a7f2-fbfba013450b-catalog-content\") pod \"community-operators-d2hs7\" (UID: \"a4a6c356-d827-4b99-a7f2-fbfba013450b\") " pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.644879 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4a6c356-d827-4b99-a7f2-fbfba013450b-utilities\") pod \"community-operators-d2hs7\" (UID: \"a4a6c356-d827-4b99-a7f2-fbfba013450b\") " pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.669195 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k5f7\" (UniqueName: \"kubernetes.io/projected/a4a6c356-d827-4b99-a7f2-fbfba013450b-kube-api-access-9k5f7\") pod \"community-operators-d2hs7\" (UID: \"a4a6c356-d827-4b99-a7f2-fbfba013450b\") " pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:45 crc kubenswrapper[5025]: I1004 10:39:45.859956 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.066379 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d2hs7"] Oct 04 10:39:46 crc kubenswrapper[5025]: W1004 10:39:46.072008 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4a6c356_d827_4b99_a7f2_fbfba013450b.slice/crio-75b92eca837751e52379ffed74fc0c63f96c677c5b096c5f7d95ec4653198c91 WatchSource:0}: Error finding container 75b92eca837751e52379ffed74fc0c63f96c677c5b096c5f7d95ec4653198c91: Status 404 returned error can't find the container with id 75b92eca837751e52379ffed74fc0c63f96c677c5b096c5f7d95ec4653198c91 Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.313763 5025 generic.go:334] "Generic (PLEG): container finished" podID="c3e7faef-f9a9-4742-b6e1-893e70ab32b3" containerID="df71a2ccf2b0e48ada5ac5b7692cc69ddeb9682e94f39bc6aa77572818674ec6" exitCode=0 Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.313846 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9hl2" event={"ID":"c3e7faef-f9a9-4742-b6e1-893e70ab32b3","Type":"ContainerDied","Data":"df71a2ccf2b0e48ada5ac5b7692cc69ddeb9682e94f39bc6aa77572818674ec6"} Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.315941 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2hs7" event={"ID":"a4a6c356-d827-4b99-a7f2-fbfba013450b","Type":"ContainerStarted","Data":"75b92eca837751e52379ffed74fc0c63f96c677c5b096c5f7d95ec4653198c91"} Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.536975 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5ln4x"] Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.538716 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.541729 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.543597 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5ln4x"] Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.554641 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5-utilities\") pod \"certified-operators-5ln4x\" (UID: \"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5\") " pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.554702 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5-catalog-content\") pod \"certified-operators-5ln4x\" (UID: \"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5\") " pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.554757 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snd64\" (UniqueName: \"kubernetes.io/projected/ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5-kube-api-access-snd64\") pod \"certified-operators-5ln4x\" (UID: \"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5\") " pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.655750 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5-utilities\") pod \"certified-operators-5ln4x\" (UID: \"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5\") " pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.655859 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5-catalog-content\") pod \"certified-operators-5ln4x\" (UID: \"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5\") " pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.655950 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snd64\" (UniqueName: \"kubernetes.io/projected/ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5-kube-api-access-snd64\") pod \"certified-operators-5ln4x\" (UID: \"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5\") " pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.656419 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5-catalog-content\") pod \"certified-operators-5ln4x\" (UID: \"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5\") " pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.656419 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5-utilities\") pod \"certified-operators-5ln4x\" (UID: \"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5\") " pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.682944 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snd64\" (UniqueName: \"kubernetes.io/projected/ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5-kube-api-access-snd64\") pod \"certified-operators-5ln4x\" (UID: \"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5\") " pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:46 crc kubenswrapper[5025]: I1004 10:39:46.852028 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:47 crc kubenswrapper[5025]: I1004 10:39:47.100209 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5ln4x"] Oct 04 10:39:47 crc kubenswrapper[5025]: W1004 10:39:47.109692 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecc4c1d1_a9f7_4cf4_b1d7_5336507676c5.slice/crio-245b9118a775f77914d3c3c2110437189d9ab5698e00c378553ff0b78dedf806 WatchSource:0}: Error finding container 245b9118a775f77914d3c3c2110437189d9ab5698e00c378553ff0b78dedf806: Status 404 returned error can't find the container with id 245b9118a775f77914d3c3c2110437189d9ab5698e00c378553ff0b78dedf806 Oct 04 10:39:47 crc kubenswrapper[5025]: I1004 10:39:47.322569 5025 generic.go:334] "Generic (PLEG): container finished" podID="ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5" containerID="b797e000250a7849e5ed28a6f4743cd3ed5fd1ffa32ad7a649893ed4b8cbb9cd" exitCode=0 Oct 04 10:39:47 crc kubenswrapper[5025]: I1004 10:39:47.322657 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ln4x" event={"ID":"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5","Type":"ContainerDied","Data":"b797e000250a7849e5ed28a6f4743cd3ed5fd1ffa32ad7a649893ed4b8cbb9cd"} Oct 04 10:39:47 crc kubenswrapper[5025]: I1004 10:39:47.322687 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ln4x" event={"ID":"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5","Type":"ContainerStarted","Data":"245b9118a775f77914d3c3c2110437189d9ab5698e00c378553ff0b78dedf806"} Oct 04 10:39:47 crc kubenswrapper[5025]: I1004 10:39:47.325206 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxpjg" event={"ID":"1b1dd0c3-486f-4ac8-be1b-cfc100de913e","Type":"ContainerStarted","Data":"ac1c53262ebd166edc5c5c3988bdc5daab69af130c9967fb5cdffffc43021c4b"} Oct 04 10:39:47 crc kubenswrapper[5025]: I1004 10:39:47.327161 5025 generic.go:334] "Generic (PLEG): container finished" podID="a4a6c356-d827-4b99-a7f2-fbfba013450b" containerID="df26cc76a719a8b7f3b5f56cf2f038db5f15a209ebc61d8c1d8f68ca61c019a4" exitCode=0 Oct 04 10:39:47 crc kubenswrapper[5025]: I1004 10:39:47.327225 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2hs7" event={"ID":"a4a6c356-d827-4b99-a7f2-fbfba013450b","Type":"ContainerDied","Data":"df26cc76a719a8b7f3b5f56cf2f038db5f15a209ebc61d8c1d8f68ca61c019a4"} Oct 04 10:39:48 crc kubenswrapper[5025]: I1004 10:39:48.334163 5025 generic.go:334] "Generic (PLEG): container finished" podID="1b1dd0c3-486f-4ac8-be1b-cfc100de913e" containerID="ac1c53262ebd166edc5c5c3988bdc5daab69af130c9967fb5cdffffc43021c4b" exitCode=0 Oct 04 10:39:48 crc kubenswrapper[5025]: I1004 10:39:48.334223 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxpjg" event={"ID":"1b1dd0c3-486f-4ac8-be1b-cfc100de913e","Type":"ContainerDied","Data":"ac1c53262ebd166edc5c5c3988bdc5daab69af130c9967fb5cdffffc43021c4b"} Oct 04 10:39:48 crc kubenswrapper[5025]: I1004 10:39:48.336972 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m9hl2" event={"ID":"c3e7faef-f9a9-4742-b6e1-893e70ab32b3","Type":"ContainerStarted","Data":"2dd3b9ef520c5b599869b7aaf3b31a7bac2f999ae271e13bc80d9baae9ee8e94"} Oct 04 10:39:48 crc kubenswrapper[5025]: I1004 10:39:48.381078 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m9hl2" podStartSLOduration=2.196424837 podStartE2EDuration="5.381056873s" podCreationTimestamp="2025-10-04 10:39:43 +0000 UTC" firstStartedPulling="2025-10-04 10:39:44.299154365 +0000 UTC m=+312.724121245" lastFinishedPulling="2025-10-04 10:39:47.483786401 +0000 UTC m=+315.908753281" observedRunningTime="2025-10-04 10:39:48.379884328 +0000 UTC m=+316.804851228" watchObservedRunningTime="2025-10-04 10:39:48.381056873 +0000 UTC m=+316.806023763" Oct 04 10:39:50 crc kubenswrapper[5025]: I1004 10:39:50.351325 5025 generic.go:334] "Generic (PLEG): container finished" podID="ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5" containerID="138f58a750861d284595afe948075d08ddf50e0cf3f6ff535814de22e0c0f56e" exitCode=0 Oct 04 10:39:50 crc kubenswrapper[5025]: I1004 10:39:50.351399 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ln4x" event={"ID":"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5","Type":"ContainerDied","Data":"138f58a750861d284595afe948075d08ddf50e0cf3f6ff535814de22e0c0f56e"} Oct 04 10:39:50 crc kubenswrapper[5025]: I1004 10:39:50.357989 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxpjg" event={"ID":"1b1dd0c3-486f-4ac8-be1b-cfc100de913e","Type":"ContainerStarted","Data":"17411c2988b8a1cf8418e32a7b19a195abe63650108b4d4b15b81cab67f8dd56"} Oct 04 10:39:50 crc kubenswrapper[5025]: I1004 10:39:50.360661 5025 generic.go:334] "Generic (PLEG): container finished" podID="a4a6c356-d827-4b99-a7f2-fbfba013450b" containerID="cedb5a24b28f0ebb3609d4f4151a576f742d720a4bc99123ad3842b32bf21e8a" exitCode=0 Oct 04 10:39:50 crc kubenswrapper[5025]: I1004 10:39:50.360690 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2hs7" event={"ID":"a4a6c356-d827-4b99-a7f2-fbfba013450b","Type":"ContainerDied","Data":"cedb5a24b28f0ebb3609d4f4151a576f742d720a4bc99123ad3842b32bf21e8a"} Oct 04 10:39:52 crc kubenswrapper[5025]: I1004 10:39:52.372134 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5ln4x" event={"ID":"ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5","Type":"ContainerStarted","Data":"1fb17711ee36324bd7b0dcc646413315f7836069ea2e6f2b453ff70c5412cbfc"} Oct 04 10:39:52 crc kubenswrapper[5025]: I1004 10:39:52.374178 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d2hs7" event={"ID":"a4a6c356-d827-4b99-a7f2-fbfba013450b","Type":"ContainerStarted","Data":"762812a60c23f956133cfb33c2b9e5f9b7144058688160d6c809b69059bac52b"} Oct 04 10:39:52 crc kubenswrapper[5025]: I1004 10:39:52.390696 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rxpjg" podStartSLOduration=4.42080462 podStartE2EDuration="8.390680066s" podCreationTimestamp="2025-10-04 10:39:44 +0000 UTC" firstStartedPulling="2025-10-04 10:39:45.434366621 +0000 UTC m=+313.859333541" lastFinishedPulling="2025-10-04 10:39:49.404242107 +0000 UTC m=+317.829208987" observedRunningTime="2025-10-04 10:39:51.396046881 +0000 UTC m=+319.821013801" watchObservedRunningTime="2025-10-04 10:39:52.390680066 +0000 UTC m=+320.815646946" Oct 04 10:39:52 crc kubenswrapper[5025]: I1004 10:39:52.392106 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5ln4x" podStartSLOduration=2.411492579 podStartE2EDuration="6.392101608s" podCreationTimestamp="2025-10-04 10:39:46 +0000 UTC" firstStartedPulling="2025-10-04 10:39:47.458447074 +0000 UTC m=+315.883413964" lastFinishedPulling="2025-10-04 10:39:51.439056113 +0000 UTC m=+319.864022993" observedRunningTime="2025-10-04 10:39:52.39012425 +0000 UTC m=+320.815091130" watchObservedRunningTime="2025-10-04 10:39:52.392101608 +0000 UTC m=+320.817068488" Oct 04 10:39:52 crc kubenswrapper[5025]: I1004 10:39:52.411671 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d2hs7" podStartSLOduration=2.817500822 podStartE2EDuration="7.411653097s" podCreationTimestamp="2025-10-04 10:39:45 +0000 UTC" firstStartedPulling="2025-10-04 10:39:47.379335851 +0000 UTC m=+315.804302731" lastFinishedPulling="2025-10-04 10:39:51.973488126 +0000 UTC m=+320.398455006" observedRunningTime="2025-10-04 10:39:52.407546367 +0000 UTC m=+320.832513257" watchObservedRunningTime="2025-10-04 10:39:52.411653097 +0000 UTC m=+320.836619977" Oct 04 10:39:53 crc kubenswrapper[5025]: I1004 10:39:53.485078 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:53 crc kubenswrapper[5025]: I1004 10:39:53.485455 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:53 crc kubenswrapper[5025]: I1004 10:39:53.533439 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:54 crc kubenswrapper[5025]: I1004 10:39:54.453564 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m9hl2" Oct 04 10:39:54 crc kubenswrapper[5025]: I1004 10:39:54.455969 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:54 crc kubenswrapper[5025]: I1004 10:39:54.456040 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:39:55 crc kubenswrapper[5025]: I1004 10:39:55.509520 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rxpjg" podUID="1b1dd0c3-486f-4ac8-be1b-cfc100de913e" containerName="registry-server" probeResult="failure" output=< Oct 04 10:39:55 crc kubenswrapper[5025]: timeout: failed to connect service ":50051" within 1s Oct 04 10:39:55 crc kubenswrapper[5025]: > Oct 04 10:39:55 crc kubenswrapper[5025]: I1004 10:39:55.861033 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:55 crc kubenswrapper[5025]: I1004 10:39:55.861107 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:55 crc kubenswrapper[5025]: I1004 10:39:55.918459 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:39:56 crc kubenswrapper[5025]: I1004 10:39:56.852944 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:56 crc kubenswrapper[5025]: I1004 10:39:56.853394 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:56 crc kubenswrapper[5025]: I1004 10:39:56.905285 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:39:57 crc kubenswrapper[5025]: I1004 10:39:57.466691 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5ln4x" Oct 04 10:40:04 crc kubenswrapper[5025]: I1004 10:40:04.523325 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:40:04 crc kubenswrapper[5025]: I1004 10:40:04.600518 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rxpjg" Oct 04 10:40:05 crc kubenswrapper[5025]: I1004 10:40:05.925839 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d2hs7" Oct 04 10:40:14 crc kubenswrapper[5025]: I1004 10:40:14.713372 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:40:14 crc kubenswrapper[5025]: I1004 10:40:14.713782 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:40:44 crc kubenswrapper[5025]: I1004 10:40:44.713572 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:40:44 crc kubenswrapper[5025]: I1004 10:40:44.714222 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:41:14 crc kubenswrapper[5025]: I1004 10:41:14.713524 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:41:14 crc kubenswrapper[5025]: I1004 10:41:14.714256 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:41:14 crc kubenswrapper[5025]: I1004 10:41:14.714315 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:41:14 crc kubenswrapper[5025]: I1004 10:41:14.715238 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db2ae36648521dea01dfdc74730ff2492700a6b36503b7ffdda74b5fe8a3c056"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:41:14 crc kubenswrapper[5025]: I1004 10:41:14.715347 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://db2ae36648521dea01dfdc74730ff2492700a6b36503b7ffdda74b5fe8a3c056" gracePeriod=600 Oct 04 10:41:14 crc kubenswrapper[5025]: I1004 10:41:14.884979 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="db2ae36648521dea01dfdc74730ff2492700a6b36503b7ffdda74b5fe8a3c056" exitCode=0 Oct 04 10:41:14 crc kubenswrapper[5025]: I1004 10:41:14.885052 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"db2ae36648521dea01dfdc74730ff2492700a6b36503b7ffdda74b5fe8a3c056"} Oct 04 10:41:14 crc kubenswrapper[5025]: I1004 10:41:14.885089 5025 scope.go:117] "RemoveContainer" containerID="fc8afcf44c9406d5081152a83f003da19aae1a43da1ac4e574cd805299ee089b" Oct 04 10:41:15 crc kubenswrapper[5025]: I1004 10:41:15.896774 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"f00d677ff24dae96198611c79163ef5caff57e489eae259b2f02ea4b187f2545"} Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.055276 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x57dl"] Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.056669 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.071217 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x57dl"] Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.088622 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2caac9e0-dcc3-4801-a10b-3e667b945437-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.088677 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2caac9e0-dcc3-4801-a10b-3e667b945437-registry-tls\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.088703 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2caac9e0-dcc3-4801-a10b-3e667b945437-bound-sa-token\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.088724 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2caac9e0-dcc3-4801-a10b-3e667b945437-trusted-ca\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.088758 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2caac9e0-dcc3-4801-a10b-3e667b945437-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.088780 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2caac9e0-dcc3-4801-a10b-3e667b945437-registry-certificates\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.088892 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.088989 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b782l\" (UniqueName: \"kubernetes.io/projected/2caac9e0-dcc3-4801-a10b-3e667b945437-kube-api-access-b782l\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.118549 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.190396 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2caac9e0-dcc3-4801-a10b-3e667b945437-registry-tls\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.190444 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2caac9e0-dcc3-4801-a10b-3e667b945437-bound-sa-token\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.190464 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2caac9e0-dcc3-4801-a10b-3e667b945437-trusted-ca\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.190502 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2caac9e0-dcc3-4801-a10b-3e667b945437-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.190534 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2caac9e0-dcc3-4801-a10b-3e667b945437-registry-certificates\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.190583 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b782l\" (UniqueName: \"kubernetes.io/projected/2caac9e0-dcc3-4801-a10b-3e667b945437-kube-api-access-b782l\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.190609 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2caac9e0-dcc3-4801-a10b-3e667b945437-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.191267 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2caac9e0-dcc3-4801-a10b-3e667b945437-ca-trust-extracted\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.191758 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2caac9e0-dcc3-4801-a10b-3e667b945437-trusted-ca\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.191854 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2caac9e0-dcc3-4801-a10b-3e667b945437-registry-certificates\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.200515 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2caac9e0-dcc3-4801-a10b-3e667b945437-installation-pull-secrets\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.201347 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2caac9e0-dcc3-4801-a10b-3e667b945437-registry-tls\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.204585 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2caac9e0-dcc3-4801-a10b-3e667b945437-bound-sa-token\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.208882 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b782l\" (UniqueName: \"kubernetes.io/projected/2caac9e0-dcc3-4801-a10b-3e667b945437-kube-api-access-b782l\") pod \"image-registry-66df7c8f76-x57dl\" (UID: \"2caac9e0-dcc3-4801-a10b-3e667b945437\") " pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.373968 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:55 crc kubenswrapper[5025]: I1004 10:42:55.593336 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-x57dl"] Oct 04 10:42:56 crc kubenswrapper[5025]: I1004 10:42:56.514603 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" event={"ID":"2caac9e0-dcc3-4801-a10b-3e667b945437","Type":"ContainerStarted","Data":"9c43d9bffc94ee8b5bc5e19c3059e375468f94c99dcbd6e24489675615c6270c"} Oct 04 10:42:56 crc kubenswrapper[5025]: I1004 10:42:56.514671 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" event={"ID":"2caac9e0-dcc3-4801-a10b-3e667b945437","Type":"ContainerStarted","Data":"a016d11229e5fa5d40a82b67fb83873ef05caacb6836144fc25ad98300ea1fb5"} Oct 04 10:42:56 crc kubenswrapper[5025]: I1004 10:42:56.514814 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:42:56 crc kubenswrapper[5025]: I1004 10:42:56.539908 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" podStartSLOduration=1.539887641 podStartE2EDuration="1.539887641s" podCreationTimestamp="2025-10-04 10:42:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:42:56.538631737 +0000 UTC m=+504.963598617" watchObservedRunningTime="2025-10-04 10:42:56.539887641 +0000 UTC m=+504.964854521" Oct 04 10:43:15 crc kubenswrapper[5025]: I1004 10:43:15.380361 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-x57dl" Oct 04 10:43:15 crc kubenswrapper[5025]: I1004 10:43:15.447194 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvzjs"] Oct 04 10:43:39 crc kubenswrapper[5025]: I1004 10:43:39.459763 5025 scope.go:117] "RemoveContainer" containerID="31be2214ed9ede5b72baef4d475d25d0c524b556cf3b5e322ddd4f38492570fb" Oct 04 10:43:40 crc kubenswrapper[5025]: I1004 10:43:40.494120 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" podUID="acc502df-2c80-40fc-975f-d993333b1898" containerName="registry" containerID="cri-o://e250e66e632078debc1e61557fcbcb5be7f1bcaa4dd4b24a3a499ffe69343f89" gracePeriod=30 Oct 04 10:43:40 crc kubenswrapper[5025]: I1004 10:43:40.783131 5025 generic.go:334] "Generic (PLEG): container finished" podID="acc502df-2c80-40fc-975f-d993333b1898" containerID="e250e66e632078debc1e61557fcbcb5be7f1bcaa4dd4b24a3a499ffe69343f89" exitCode=0 Oct 04 10:43:40 crc kubenswrapper[5025]: I1004 10:43:40.783541 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" event={"ID":"acc502df-2c80-40fc-975f-d993333b1898","Type":"ContainerDied","Data":"e250e66e632078debc1e61557fcbcb5be7f1bcaa4dd4b24a3a499ffe69343f89"} Oct 04 10:43:40 crc kubenswrapper[5025]: I1004 10:43:40.849082 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.033799 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/acc502df-2c80-40fc-975f-d993333b1898-installation-pull-secrets\") pod \"acc502df-2c80-40fc-975f-d993333b1898\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.033893 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/acc502df-2c80-40fc-975f-d993333b1898-ca-trust-extracted\") pod \"acc502df-2c80-40fc-975f-d993333b1898\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.033926 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-trusted-ca\") pod \"acc502df-2c80-40fc-975f-d993333b1898\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.034007 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-registry-tls\") pod \"acc502df-2c80-40fc-975f-d993333b1898\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.034088 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-registry-certificates\") pod \"acc502df-2c80-40fc-975f-d993333b1898\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.034105 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltskc\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-kube-api-access-ltskc\") pod \"acc502df-2c80-40fc-975f-d993333b1898\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.034144 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-bound-sa-token\") pod \"acc502df-2c80-40fc-975f-d993333b1898\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.034311 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"acc502df-2c80-40fc-975f-d993333b1898\" (UID: \"acc502df-2c80-40fc-975f-d993333b1898\") " Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.035252 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "acc502df-2c80-40fc-975f-d993333b1898" (UID: "acc502df-2c80-40fc-975f-d993333b1898"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.035467 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "acc502df-2c80-40fc-975f-d993333b1898" (UID: "acc502df-2c80-40fc-975f-d993333b1898"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.040101 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acc502df-2c80-40fc-975f-d993333b1898-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "acc502df-2c80-40fc-975f-d993333b1898" (UID: "acc502df-2c80-40fc-975f-d993333b1898"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.040965 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "acc502df-2c80-40fc-975f-d993333b1898" (UID: "acc502df-2c80-40fc-975f-d993333b1898"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.041291 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "acc502df-2c80-40fc-975f-d993333b1898" (UID: "acc502df-2c80-40fc-975f-d993333b1898"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.041497 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-kube-api-access-ltskc" (OuterVolumeSpecName: "kube-api-access-ltskc") pod "acc502df-2c80-40fc-975f-d993333b1898" (UID: "acc502df-2c80-40fc-975f-d993333b1898"). InnerVolumeSpecName "kube-api-access-ltskc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.049355 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "acc502df-2c80-40fc-975f-d993333b1898" (UID: "acc502df-2c80-40fc-975f-d993333b1898"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.050673 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acc502df-2c80-40fc-975f-d993333b1898-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "acc502df-2c80-40fc-975f-d993333b1898" (UID: "acc502df-2c80-40fc-975f-d993333b1898"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.135468 5025 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.135532 5025 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.135555 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltskc\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-kube-api-access-ltskc\") on node \"crc\" DevicePath \"\"" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.135574 5025 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/acc502df-2c80-40fc-975f-d993333b1898-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.135594 5025 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/acc502df-2c80-40fc-975f-d993333b1898-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.135615 5025 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/acc502df-2c80-40fc-975f-d993333b1898-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.135633 5025 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/acc502df-2c80-40fc-975f-d993333b1898-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.791985 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" event={"ID":"acc502df-2c80-40fc-975f-d993333b1898","Type":"ContainerDied","Data":"f1e7fe83b58920d35473b32f01edf9e876904fe8bc4b9bdcd39d8f64d2235fcf"} Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.792113 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pvzjs" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.792129 5025 scope.go:117] "RemoveContainer" containerID="e250e66e632078debc1e61557fcbcb5be7f1bcaa4dd4b24a3a499ffe69343f89" Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.835929 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvzjs"] Oct 04 10:43:41 crc kubenswrapper[5025]: I1004 10:43:41.843144 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvzjs"] Oct 04 10:43:42 crc kubenswrapper[5025]: I1004 10:43:42.421902 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acc502df-2c80-40fc-975f-d993333b1898" path="/var/lib/kubelet/pods/acc502df-2c80-40fc-975f-d993333b1898/volumes" Oct 04 10:43:44 crc kubenswrapper[5025]: I1004 10:43:44.713315 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:43:44 crc kubenswrapper[5025]: I1004 10:43:44.713761 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:44:14 crc kubenswrapper[5025]: I1004 10:44:14.713941 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:44:14 crc kubenswrapper[5025]: I1004 10:44:14.714791 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:44:44 crc kubenswrapper[5025]: I1004 10:44:44.713650 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:44:44 crc kubenswrapper[5025]: I1004 10:44:44.714192 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:44:44 crc kubenswrapper[5025]: I1004 10:44:44.714257 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:44:44 crc kubenswrapper[5025]: I1004 10:44:44.715202 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f00d677ff24dae96198611c79163ef5caff57e489eae259b2f02ea4b187f2545"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:44:44 crc kubenswrapper[5025]: I1004 10:44:44.715308 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://f00d677ff24dae96198611c79163ef5caff57e489eae259b2f02ea4b187f2545" gracePeriod=600 Oct 04 10:44:45 crc kubenswrapper[5025]: I1004 10:44:45.189917 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="f00d677ff24dae96198611c79163ef5caff57e489eae259b2f02ea4b187f2545" exitCode=0 Oct 04 10:44:45 crc kubenswrapper[5025]: I1004 10:44:45.189975 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"f00d677ff24dae96198611c79163ef5caff57e489eae259b2f02ea4b187f2545"} Oct 04 10:44:45 crc kubenswrapper[5025]: I1004 10:44:45.190042 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"4ca2d393da7fa0a921a67c473465c531835d20e5d5778a21a26fa2e7ee5204ae"} Oct 04 10:44:45 crc kubenswrapper[5025]: I1004 10:44:45.190072 5025 scope.go:117] "RemoveContainer" containerID="db2ae36648521dea01dfdc74730ff2492700a6b36503b7ffdda74b5fe8a3c056" Oct 04 10:44:55 crc kubenswrapper[5025]: I1004 10:44:55.994820 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-54hps"] Oct 04 10:44:55 crc kubenswrapper[5025]: E1004 10:44:55.995635 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acc502df-2c80-40fc-975f-d993333b1898" containerName="registry" Oct 04 10:44:55 crc kubenswrapper[5025]: I1004 10:44:55.995650 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="acc502df-2c80-40fc-975f-d993333b1898" containerName="registry" Oct 04 10:44:55 crc kubenswrapper[5025]: I1004 10:44:55.995765 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="acc502df-2c80-40fc-975f-d993333b1898" containerName="registry" Oct 04 10:44:55 crc kubenswrapper[5025]: I1004 10:44:55.996210 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-54hps" Oct 04 10:44:55 crc kubenswrapper[5025]: I1004 10:44:55.997614 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 04 10:44:55 crc kubenswrapper[5025]: I1004 10:44:55.997888 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 04 10:44:55 crc kubenswrapper[5025]: I1004 10:44:55.998161 5025 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-svhsq" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.006739 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-54hps"] Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.016970 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hxgr5"] Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.017812 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-hxgr5" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.019939 5025 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-p7p8d" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.025120 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8h2xb"] Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.025882 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-8h2xb" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.027544 5025 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-gsfk8" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.040800 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8h2xb"] Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.066001 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hxgr5"] Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.148724 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbsgp\" (UniqueName: \"kubernetes.io/projected/8f18ef59-714b-4044-b76f-44cb4b523a3f-kube-api-access-mbsgp\") pod \"cert-manager-5b446d88c5-hxgr5\" (UID: \"8f18ef59-714b-4044-b76f-44cb4b523a3f\") " pod="cert-manager/cert-manager-5b446d88c5-hxgr5" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.148784 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n4nt\" (UniqueName: \"kubernetes.io/projected/a413e9a9-6200-4e4a-abb7-939ed023a70f-kube-api-access-9n4nt\") pod \"cert-manager-cainjector-7f985d654d-54hps\" (UID: \"a413e9a9-6200-4e4a-abb7-939ed023a70f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-54hps" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.148804 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j529p\" (UniqueName: \"kubernetes.io/projected/a45d9552-c070-4b0b-99b7-50cfd7acbd28-kube-api-access-j529p\") pod \"cert-manager-webhook-5655c58dd6-8h2xb\" (UID: \"a45d9552-c070-4b0b-99b7-50cfd7acbd28\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8h2xb" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.249420 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbsgp\" (UniqueName: \"kubernetes.io/projected/8f18ef59-714b-4044-b76f-44cb4b523a3f-kube-api-access-mbsgp\") pod \"cert-manager-5b446d88c5-hxgr5\" (UID: \"8f18ef59-714b-4044-b76f-44cb4b523a3f\") " pod="cert-manager/cert-manager-5b446d88c5-hxgr5" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.249488 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n4nt\" (UniqueName: \"kubernetes.io/projected/a413e9a9-6200-4e4a-abb7-939ed023a70f-kube-api-access-9n4nt\") pod \"cert-manager-cainjector-7f985d654d-54hps\" (UID: \"a413e9a9-6200-4e4a-abb7-939ed023a70f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-54hps" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.249517 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j529p\" (UniqueName: \"kubernetes.io/projected/a45d9552-c070-4b0b-99b7-50cfd7acbd28-kube-api-access-j529p\") pod \"cert-manager-webhook-5655c58dd6-8h2xb\" (UID: \"a45d9552-c070-4b0b-99b7-50cfd7acbd28\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8h2xb" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.269406 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j529p\" (UniqueName: \"kubernetes.io/projected/a45d9552-c070-4b0b-99b7-50cfd7acbd28-kube-api-access-j529p\") pod \"cert-manager-webhook-5655c58dd6-8h2xb\" (UID: \"a45d9552-c070-4b0b-99b7-50cfd7acbd28\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8h2xb" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.271002 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbsgp\" (UniqueName: \"kubernetes.io/projected/8f18ef59-714b-4044-b76f-44cb4b523a3f-kube-api-access-mbsgp\") pod \"cert-manager-5b446d88c5-hxgr5\" (UID: \"8f18ef59-714b-4044-b76f-44cb4b523a3f\") " pod="cert-manager/cert-manager-5b446d88c5-hxgr5" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.271918 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n4nt\" (UniqueName: \"kubernetes.io/projected/a413e9a9-6200-4e4a-abb7-939ed023a70f-kube-api-access-9n4nt\") pod \"cert-manager-cainjector-7f985d654d-54hps\" (UID: \"a413e9a9-6200-4e4a-abb7-939ed023a70f\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-54hps" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.310120 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-54hps" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.331210 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-hxgr5" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.340383 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-8h2xb" Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.545483 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-54hps"] Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.554648 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.790508 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8h2xb"] Oct 04 10:44:56 crc kubenswrapper[5025]: I1004 10:44:56.801577 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hxgr5"] Oct 04 10:44:57 crc kubenswrapper[5025]: I1004 10:44:57.259907 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-54hps" event={"ID":"a413e9a9-6200-4e4a-abb7-939ed023a70f","Type":"ContainerStarted","Data":"15a728814d5fee255335b5d95a9342e8a5c8bf33de9ac97c3190fc287571f7cb"} Oct 04 10:44:57 crc kubenswrapper[5025]: I1004 10:44:57.261275 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-hxgr5" event={"ID":"8f18ef59-714b-4044-b76f-44cb4b523a3f","Type":"ContainerStarted","Data":"bb5a57e4ab61b20713ff16966b2c4511075977fd48e1a74bf94b5bdf4f34b348"} Oct 04 10:44:57 crc kubenswrapper[5025]: I1004 10:44:57.262409 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-8h2xb" event={"ID":"a45d9552-c070-4b0b-99b7-50cfd7acbd28","Type":"ContainerStarted","Data":"33e82db9020a4ce20af311338cc443a84cc360a074d66ee56bce2407c73eee26"} Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.123996 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk"] Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.125337 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.127050 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.133624 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.134552 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk"] Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.299317 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-config-volume\") pod \"collect-profiles-29326245-glndk\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.299580 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xldzl\" (UniqueName: \"kubernetes.io/projected/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-kube-api-access-xldzl\") pod \"collect-profiles-29326245-glndk\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.300065 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-secret-volume\") pod \"collect-profiles-29326245-glndk\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.401046 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-secret-volume\") pod \"collect-profiles-29326245-glndk\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.401144 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-config-volume\") pod \"collect-profiles-29326245-glndk\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.401175 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xldzl\" (UniqueName: \"kubernetes.io/projected/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-kube-api-access-xldzl\") pod \"collect-profiles-29326245-glndk\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.402521 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-config-volume\") pod \"collect-profiles-29326245-glndk\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.406514 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-secret-volume\") pod \"collect-profiles-29326245-glndk\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.416717 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xldzl\" (UniqueName: \"kubernetes.io/projected/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-kube-api-access-xldzl\") pod \"collect-profiles-29326245-glndk\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.447485 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:00 crc kubenswrapper[5025]: I1004 10:45:00.964345 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk"] Oct 04 10:45:00 crc kubenswrapper[5025]: W1004 10:45:00.972003 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96fc8ecd_3f6e_4bd6_ad7c_6ca2145883a8.slice/crio-9e2dee039e226543e56b7c891baea3966c8f485fda80a1cb0dde8cdd34a78987 WatchSource:0}: Error finding container 9e2dee039e226543e56b7c891baea3966c8f485fda80a1cb0dde8cdd34a78987: Status 404 returned error can't find the container with id 9e2dee039e226543e56b7c891baea3966c8f485fda80a1cb0dde8cdd34a78987 Oct 04 10:45:01 crc kubenswrapper[5025]: I1004 10:45:01.284296 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-8h2xb" event={"ID":"a45d9552-c070-4b0b-99b7-50cfd7acbd28","Type":"ContainerStarted","Data":"ffacc80657303630687ea3665df99216b72957287db469ca3022db55b240a5de"} Oct 04 10:45:01 crc kubenswrapper[5025]: I1004 10:45:01.285049 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-8h2xb" Oct 04 10:45:01 crc kubenswrapper[5025]: I1004 10:45:01.286519 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-54hps" event={"ID":"a413e9a9-6200-4e4a-abb7-939ed023a70f","Type":"ContainerStarted","Data":"1b96ce230caecc2723d94a9e2a61dcb47fee8b4b4d821e85964fc1aa4193d657"} Oct 04 10:45:01 crc kubenswrapper[5025]: I1004 10:45:01.287811 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-hxgr5" event={"ID":"8f18ef59-714b-4044-b76f-44cb4b523a3f","Type":"ContainerStarted","Data":"57da6d35d6f0146af1ee0816f225589990eefb639d975fef28b39bd65138f7d1"} Oct 04 10:45:01 crc kubenswrapper[5025]: I1004 10:45:01.289075 5025 generic.go:334] "Generic (PLEG): container finished" podID="96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8" containerID="11aac83d485b7ae8776577868850b293b8daa24f8cb8b1da3f1fc992e40c2afb" exitCode=0 Oct 04 10:45:01 crc kubenswrapper[5025]: I1004 10:45:01.289110 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" event={"ID":"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8","Type":"ContainerDied","Data":"11aac83d485b7ae8776577868850b293b8daa24f8cb8b1da3f1fc992e40c2afb"} Oct 04 10:45:01 crc kubenswrapper[5025]: I1004 10:45:01.289130 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" event={"ID":"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8","Type":"ContainerStarted","Data":"9e2dee039e226543e56b7c891baea3966c8f485fda80a1cb0dde8cdd34a78987"} Oct 04 10:45:01 crc kubenswrapper[5025]: I1004 10:45:01.302827 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-8h2xb" podStartSLOduration=1.403020385 podStartE2EDuration="5.302810498s" podCreationTimestamp="2025-10-04 10:44:56 +0000 UTC" firstStartedPulling="2025-10-04 10:44:56.801411064 +0000 UTC m=+625.226377944" lastFinishedPulling="2025-10-04 10:45:00.701201167 +0000 UTC m=+629.126168057" observedRunningTime="2025-10-04 10:45:01.29981834 +0000 UTC m=+629.724785230" watchObservedRunningTime="2025-10-04 10:45:01.302810498 +0000 UTC m=+629.727777378" Oct 04 10:45:01 crc kubenswrapper[5025]: I1004 10:45:01.314563 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-54hps" podStartSLOduration=2.174235477 podStartE2EDuration="6.314541714s" podCreationTimestamp="2025-10-04 10:44:55 +0000 UTC" firstStartedPulling="2025-10-04 10:44:56.554284405 +0000 UTC m=+624.979251285" lastFinishedPulling="2025-10-04 10:45:00.694590612 +0000 UTC m=+629.119557522" observedRunningTime="2025-10-04 10:45:01.313092412 +0000 UTC m=+629.738059292" watchObservedRunningTime="2025-10-04 10:45:01.314541714 +0000 UTC m=+629.739508594" Oct 04 10:45:01 crc kubenswrapper[5025]: I1004 10:45:01.332849 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-hxgr5" podStartSLOduration=2.432156765 podStartE2EDuration="6.332832134s" podCreationTimestamp="2025-10-04 10:44:55 +0000 UTC" firstStartedPulling="2025-10-04 10:44:56.80706481 +0000 UTC m=+625.232031690" lastFinishedPulling="2025-10-04 10:45:00.707740169 +0000 UTC m=+629.132707059" observedRunningTime="2025-10-04 10:45:01.331957658 +0000 UTC m=+629.756924538" watchObservedRunningTime="2025-10-04 10:45:01.332832134 +0000 UTC m=+629.757799014" Oct 04 10:45:02 crc kubenswrapper[5025]: I1004 10:45:02.564239 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:02 crc kubenswrapper[5025]: I1004 10:45:02.648868 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-secret-volume\") pod \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " Oct 04 10:45:02 crc kubenswrapper[5025]: I1004 10:45:02.649736 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-config-volume" (OuterVolumeSpecName: "config-volume") pod "96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8" (UID: "96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:45:02 crc kubenswrapper[5025]: I1004 10:45:02.649797 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-config-volume\") pod \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " Oct 04 10:45:02 crc kubenswrapper[5025]: I1004 10:45:02.649933 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xldzl\" (UniqueName: \"kubernetes.io/projected/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-kube-api-access-xldzl\") pod \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\" (UID: \"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8\") " Oct 04 10:45:02 crc kubenswrapper[5025]: I1004 10:45:02.650862 5025 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:02 crc kubenswrapper[5025]: I1004 10:45:02.656662 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8" (UID: "96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:45:02 crc kubenswrapper[5025]: I1004 10:45:02.656672 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-kube-api-access-xldzl" (OuterVolumeSpecName: "kube-api-access-xldzl") pod "96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8" (UID: "96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8"). InnerVolumeSpecName "kube-api-access-xldzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:45:02 crc kubenswrapper[5025]: I1004 10:45:02.751737 5025 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:02 crc kubenswrapper[5025]: I1004 10:45:02.751780 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xldzl\" (UniqueName: \"kubernetes.io/projected/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8-kube-api-access-xldzl\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:03 crc kubenswrapper[5025]: I1004 10:45:03.306235 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" event={"ID":"96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8","Type":"ContainerDied","Data":"9e2dee039e226543e56b7c891baea3966c8f485fda80a1cb0dde8cdd34a78987"} Oct 04 10:45:03 crc kubenswrapper[5025]: I1004 10:45:03.306267 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk" Oct 04 10:45:03 crc kubenswrapper[5025]: I1004 10:45:03.306289 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e2dee039e226543e56b7c891baea3966c8f485fda80a1cb0dde8cdd34a78987" Oct 04 10:45:06 crc kubenswrapper[5025]: I1004 10:45:06.345087 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-8h2xb" Oct 04 10:45:06 crc kubenswrapper[5025]: I1004 10:45:06.816857 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bsdl9"] Oct 04 10:45:06 crc kubenswrapper[5025]: I1004 10:45:06.817440 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovn-controller" containerID="cri-o://9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae" gracePeriod=30 Oct 04 10:45:06 crc kubenswrapper[5025]: I1004 10:45:06.817522 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="nbdb" containerID="cri-o://a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4" gracePeriod=30 Oct 04 10:45:06 crc kubenswrapper[5025]: I1004 10:45:06.817593 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="northd" containerID="cri-o://a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648" gracePeriod=30 Oct 04 10:45:06 crc kubenswrapper[5025]: I1004 10:45:06.817656 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4" gracePeriod=30 Oct 04 10:45:06 crc kubenswrapper[5025]: I1004 10:45:06.817707 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="kube-rbac-proxy-node" containerID="cri-o://7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906" gracePeriod=30 Oct 04 10:45:06 crc kubenswrapper[5025]: I1004 10:45:06.817756 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovn-acl-logging" containerID="cri-o://b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95" gracePeriod=30 Oct 04 10:45:06 crc kubenswrapper[5025]: I1004 10:45:06.817891 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="sbdb" containerID="cri-o://80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9" gracePeriod=30 Oct 04 10:45:06 crc kubenswrapper[5025]: I1004 10:45:06.871801 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" containerID="cri-o://01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a" gracePeriod=30 Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.163758 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/3.log" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.166075 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovn-acl-logging/0.log" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.166536 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovn-controller/0.log" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.166904 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.210659 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-systemd-units\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.210805 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-etc-openvswitch\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.210962 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-node-log\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211107 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z75f4\" (UniqueName: \"kubernetes.io/projected/607a1d66-62fc-4dba-9c44-6798f087fb5c-kube-api-access-z75f4\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211113 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211121 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211137 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-bin\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211204 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-node-log" (OuterVolumeSpecName: "node-log") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211228 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-env-overrides\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211297 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211312 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-slash\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211365 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-ovn-kubernetes\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211420 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-script-lib\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211445 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-slash" (OuterVolumeSpecName: "host-slash") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211506 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-kubelet\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211583 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211668 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211737 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-systemd\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211765 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-openvswitch\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211719 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211815 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211963 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-ovn\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211887 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.211884 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212095 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-netd\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212162 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-netns\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212151 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212176 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212229 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-log-socket\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212253 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212287 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-log-socket" (OuterVolumeSpecName: "log-socket") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212296 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-var-lib-openvswitch\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212318 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212365 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-config\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212400 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovn-node-metrics-cert\") pod \"607a1d66-62fc-4dba-9c44-6798f087fb5c\" (UID: \"607a1d66-62fc-4dba-9c44-6798f087fb5c\") " Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.212743 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213463 5025 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213482 5025 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213491 5025 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-slash\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213500 5025 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213510 5025 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213518 5025 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213526 5025 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213535 5025 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213545 5025 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213553 5025 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213561 5025 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-log-socket\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213568 5025 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213576 5025 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213585 5025 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213593 5025 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.213600 5025 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-node-log\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.215731 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.217549 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/607a1d66-62fc-4dba-9c44-6798f087fb5c-kube-api-access-z75f4" (OuterVolumeSpecName: "kube-api-access-z75f4") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "kube-api-access-z75f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.218498 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.219617 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zmq6r"] Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.219832 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="kube-rbac-proxy-node" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.219852 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="kube-rbac-proxy-node" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.219864 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="nbdb" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.219872 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="nbdb" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.219883 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.219890 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.219899 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.219905 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.219915 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovn-acl-logging" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.219922 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovn-acl-logging" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.219932 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="northd" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.219939 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="northd" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.219951 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8" containerName="collect-profiles" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.219960 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8" containerName="collect-profiles" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.219971 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovn-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.219978 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovn-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.219989 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.219997 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.220007 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220030 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.220044 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="sbdb" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220052 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="sbdb" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.220064 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="kubecfg-setup" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220074 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="kubecfg-setup" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220183 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220196 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8" containerName="collect-profiles" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220206 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220215 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220223 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="nbdb" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220229 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220237 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="kube-rbac-proxy-ovn-metrics" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220249 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="northd" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220257 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="sbdb" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220264 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="kube-rbac-proxy-node" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220273 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovn-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220282 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovn-acl-logging" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.220432 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220443 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.220457 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220465 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.220644 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerName="ovnkube-controller" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.224147 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.228417 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "607a1d66-62fc-4dba-9c44-6798f087fb5c" (UID: "607a1d66-62fc-4dba-9c44-6798f087fb5c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.314410 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-run-ovn-kubernetes\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.314847 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tb4p\" (UniqueName: \"kubernetes.io/projected/12c41f75-e5d5-47e3-a4bb-162c922a5912-kube-api-access-7tb4p\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.314927 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-cni-netd\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.314968 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-log-socket\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315055 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-slash\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315085 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315132 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-run-ovn\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315155 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/12c41f75-e5d5-47e3-a4bb-162c922a5912-ovnkube-script-lib\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315250 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/12c41f75-e5d5-47e3-a4bb-162c922a5912-ovnkube-config\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315299 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-var-lib-openvswitch\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315323 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/12c41f75-e5d5-47e3-a4bb-162c922a5912-ovn-node-metrics-cert\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315356 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-kubelet\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315485 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-run-systemd\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315566 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-run-netns\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315599 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-cni-bin\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315650 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/12c41f75-e5d5-47e3-a4bb-162c922a5912-env-overrides\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315695 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-run-openvswitch\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315729 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-systemd-units\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315778 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-node-log\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315810 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-etc-openvswitch\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315888 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z75f4\" (UniqueName: \"kubernetes.io/projected/607a1d66-62fc-4dba-9c44-6798f087fb5c-kube-api-access-z75f4\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315908 5025 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315922 5025 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/607a1d66-62fc-4dba-9c44-6798f087fb5c-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.315936 5025 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/607a1d66-62fc-4dba-9c44-6798f087fb5c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.331454 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n49fv_8d214492-7532-4f72-b032-b3cc1ae19473/kube-multus/2.log" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.332296 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n49fv_8d214492-7532-4f72-b032-b3cc1ae19473/kube-multus/1.log" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.332355 5025 generic.go:334] "Generic (PLEG): container finished" podID="8d214492-7532-4f72-b032-b3cc1ae19473" containerID="f1d8ee42e06786ad69717a96ff5cdc23094d6c2bb0f8a56b8a35ed098ba52011" exitCode=2 Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.332434 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n49fv" event={"ID":"8d214492-7532-4f72-b032-b3cc1ae19473","Type":"ContainerDied","Data":"f1d8ee42e06786ad69717a96ff5cdc23094d6c2bb0f8a56b8a35ed098ba52011"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.332491 5025 scope.go:117] "RemoveContainer" containerID="30cdba575bb82b7075cc835eb8c84bd61e6026e60132d02ea8d7c41143b98850" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.333339 5025 scope.go:117] "RemoveContainer" containerID="f1d8ee42e06786ad69717a96ff5cdc23094d6c2bb0f8a56b8a35ed098ba52011" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.334174 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-n49fv_openshift-multus(8d214492-7532-4f72-b032-b3cc1ae19473)\"" pod="openshift-multus/multus-n49fv" podUID="8d214492-7532-4f72-b032-b3cc1ae19473" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.335297 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovnkube-controller/3.log" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.337535 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovn-acl-logging/0.log" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338200 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bsdl9_607a1d66-62fc-4dba-9c44-6798f087fb5c/ovn-controller/0.log" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338767 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a" exitCode=0 Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338788 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9" exitCode=0 Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338798 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4" exitCode=0 Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338807 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648" exitCode=0 Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338816 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4" exitCode=0 Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338824 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906" exitCode=0 Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338833 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95" exitCode=143 Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338843 5025 generic.go:334] "Generic (PLEG): container finished" podID="607a1d66-62fc-4dba-9c44-6798f087fb5c" containerID="9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae" exitCode=143 Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338864 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338890 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338904 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338918 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338931 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338945 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338959 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338965 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.338971 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339038 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339046 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339053 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339059 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339066 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339072 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339078 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339085 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339095 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339107 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339114 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339120 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339126 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339132 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339138 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339144 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339150 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339157 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339163 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339172 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339183 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339191 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339198 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339204 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339210 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339217 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339223 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339230 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339238 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339246 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339256 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bsdl9" event={"ID":"607a1d66-62fc-4dba-9c44-6798f087fb5c","Type":"ContainerDied","Data":"9e7005ffd447d70030f29b8d24ec07b4986c5d1cde8ec69703fc2a3907044941"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339268 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339275 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339282 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339290 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339296 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339303 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339310 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339317 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339323 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.339329 5025 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495"} Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.362072 5025 scope.go:117] "RemoveContainer" containerID="01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.381816 5025 scope.go:117] "RemoveContainer" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.384853 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bsdl9"] Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.388602 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bsdl9"] Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.401698 5025 scope.go:117] "RemoveContainer" containerID="80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.411612 5025 scope.go:117] "RemoveContainer" containerID="a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417250 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-kubelet\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417279 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-run-systemd\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417298 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-cni-bin\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417314 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-run-netns\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417335 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/12c41f75-e5d5-47e3-a4bb-162c922a5912-env-overrides\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417346 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-kubelet\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417357 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-run-openvswitch\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417374 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-cni-bin\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417377 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-run-systemd\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417423 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-run-netns\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417426 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-run-openvswitch\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417402 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-systemd-units\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417381 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-systemd-units\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417473 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-node-log\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417493 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-etc-openvswitch\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417523 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-run-ovn-kubernetes\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417534 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-node-log\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417556 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-etc-openvswitch\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417565 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tb4p\" (UniqueName: \"kubernetes.io/projected/12c41f75-e5d5-47e3-a4bb-162c922a5912-kube-api-access-7tb4p\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417577 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-run-ovn-kubernetes\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417589 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-cni-netd\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417615 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-log-socket\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417648 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-slash\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417683 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417710 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-run-ovn\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417730 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/12c41f75-e5d5-47e3-a4bb-162c922a5912-ovnkube-script-lib\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417774 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/12c41f75-e5d5-47e3-a4bb-162c922a5912-env-overrides\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417780 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-var-lib-openvswitch\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417801 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/12c41f75-e5d5-47e3-a4bb-162c922a5912-ovnkube-config\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417809 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417824 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/12c41f75-e5d5-47e3-a4bb-162c922a5912-ovn-node-metrics-cert\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417831 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-cni-netd\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.417893 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-run-ovn\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.418174 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-host-slash\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.418264 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-log-socket\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.418507 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/12c41f75-e5d5-47e3-a4bb-162c922a5912-ovnkube-script-lib\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.418701 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/12c41f75-e5d5-47e3-a4bb-162c922a5912-ovnkube-config\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.418806 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/12c41f75-e5d5-47e3-a4bb-162c922a5912-var-lib-openvswitch\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.421200 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/12c41f75-e5d5-47e3-a4bb-162c922a5912-ovn-node-metrics-cert\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.425348 5025 scope.go:117] "RemoveContainer" containerID="a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.433315 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tb4p\" (UniqueName: \"kubernetes.io/projected/12c41f75-e5d5-47e3-a4bb-162c922a5912-kube-api-access-7tb4p\") pod \"ovnkube-node-zmq6r\" (UID: \"12c41f75-e5d5-47e3-a4bb-162c922a5912\") " pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.439546 5025 scope.go:117] "RemoveContainer" containerID="8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.451493 5025 scope.go:117] "RemoveContainer" containerID="7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.462062 5025 scope.go:117] "RemoveContainer" containerID="b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.472982 5025 scope.go:117] "RemoveContainer" containerID="9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.486029 5025 scope.go:117] "RemoveContainer" containerID="e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.497528 5025 scope.go:117] "RemoveContainer" containerID="01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.497951 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a\": container with ID starting with 01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a not found: ID does not exist" containerID="01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.497987 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a"} err="failed to get container status \"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a\": rpc error: code = NotFound desc = could not find container \"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a\": container with ID starting with 01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.498025 5025 scope.go:117] "RemoveContainer" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.498377 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\": container with ID starting with a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2 not found: ID does not exist" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.498440 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2"} err="failed to get container status \"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\": rpc error: code = NotFound desc = could not find container \"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\": container with ID starting with a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.498474 5025 scope.go:117] "RemoveContainer" containerID="80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.498779 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\": container with ID starting with 80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9 not found: ID does not exist" containerID="80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.498806 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9"} err="failed to get container status \"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\": rpc error: code = NotFound desc = could not find container \"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\": container with ID starting with 80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.498821 5025 scope.go:117] "RemoveContainer" containerID="a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.499093 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\": container with ID starting with a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4 not found: ID does not exist" containerID="a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.499123 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4"} err="failed to get container status \"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\": rpc error: code = NotFound desc = could not find container \"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\": container with ID starting with a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.499141 5025 scope.go:117] "RemoveContainer" containerID="a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.499364 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\": container with ID starting with a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648 not found: ID does not exist" containerID="a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.499387 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648"} err="failed to get container status \"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\": rpc error: code = NotFound desc = could not find container \"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\": container with ID starting with a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.499402 5025 scope.go:117] "RemoveContainer" containerID="8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.499577 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\": container with ID starting with 8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4 not found: ID does not exist" containerID="8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.499603 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4"} err="failed to get container status \"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\": rpc error: code = NotFound desc = could not find container \"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\": container with ID starting with 8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.499619 5025 scope.go:117] "RemoveContainer" containerID="7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.499842 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\": container with ID starting with 7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906 not found: ID does not exist" containerID="7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.499861 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906"} err="failed to get container status \"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\": rpc error: code = NotFound desc = could not find container \"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\": container with ID starting with 7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.499872 5025 scope.go:117] "RemoveContainer" containerID="b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.500175 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\": container with ID starting with b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95 not found: ID does not exist" containerID="b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.500210 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95"} err="failed to get container status \"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\": rpc error: code = NotFound desc = could not find container \"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\": container with ID starting with b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.500222 5025 scope.go:117] "RemoveContainer" containerID="9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.500639 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\": container with ID starting with 9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae not found: ID does not exist" containerID="9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.500660 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae"} err="failed to get container status \"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\": rpc error: code = NotFound desc = could not find container \"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\": container with ID starting with 9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.500672 5025 scope.go:117] "RemoveContainer" containerID="e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495" Oct 04 10:45:07 crc kubenswrapper[5025]: E1004 10:45:07.500888 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\": container with ID starting with e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495 not found: ID does not exist" containerID="e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.500911 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495"} err="failed to get container status \"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\": rpc error: code = NotFound desc = could not find container \"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\": container with ID starting with e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.500930 5025 scope.go:117] "RemoveContainer" containerID="01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.501225 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a"} err="failed to get container status \"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a\": rpc error: code = NotFound desc = could not find container \"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a\": container with ID starting with 01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.501253 5025 scope.go:117] "RemoveContainer" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.501462 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2"} err="failed to get container status \"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\": rpc error: code = NotFound desc = could not find container \"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\": container with ID starting with a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.501488 5025 scope.go:117] "RemoveContainer" containerID="80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.501688 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9"} err="failed to get container status \"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\": rpc error: code = NotFound desc = could not find container \"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\": container with ID starting with 80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.501714 5025 scope.go:117] "RemoveContainer" containerID="a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.501930 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4"} err="failed to get container status \"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\": rpc error: code = NotFound desc = could not find container \"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\": container with ID starting with a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.501962 5025 scope.go:117] "RemoveContainer" containerID="a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.502232 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648"} err="failed to get container status \"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\": rpc error: code = NotFound desc = could not find container \"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\": container with ID starting with a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.502258 5025 scope.go:117] "RemoveContainer" containerID="8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.502461 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4"} err="failed to get container status \"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\": rpc error: code = NotFound desc = could not find container \"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\": container with ID starting with 8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.502484 5025 scope.go:117] "RemoveContainer" containerID="7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.502775 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906"} err="failed to get container status \"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\": rpc error: code = NotFound desc = could not find container \"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\": container with ID starting with 7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.502809 5025 scope.go:117] "RemoveContainer" containerID="b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.503002 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95"} err="failed to get container status \"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\": rpc error: code = NotFound desc = could not find container \"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\": container with ID starting with b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.503039 5025 scope.go:117] "RemoveContainer" containerID="9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.503293 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae"} err="failed to get container status \"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\": rpc error: code = NotFound desc = could not find container \"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\": container with ID starting with 9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.503312 5025 scope.go:117] "RemoveContainer" containerID="e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.503492 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495"} err="failed to get container status \"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\": rpc error: code = NotFound desc = could not find container \"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\": container with ID starting with e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.503515 5025 scope.go:117] "RemoveContainer" containerID="01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.503674 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a"} err="failed to get container status \"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a\": rpc error: code = NotFound desc = could not find container \"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a\": container with ID starting with 01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.503699 5025 scope.go:117] "RemoveContainer" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.503884 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2"} err="failed to get container status \"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\": rpc error: code = NotFound desc = could not find container \"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\": container with ID starting with a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.503936 5025 scope.go:117] "RemoveContainer" containerID="80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.504185 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9"} err="failed to get container status \"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\": rpc error: code = NotFound desc = could not find container \"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\": container with ID starting with 80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.504208 5025 scope.go:117] "RemoveContainer" containerID="a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.504452 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4"} err="failed to get container status \"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\": rpc error: code = NotFound desc = could not find container \"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\": container with ID starting with a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.504471 5025 scope.go:117] "RemoveContainer" containerID="a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.504725 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648"} err="failed to get container status \"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\": rpc error: code = NotFound desc = could not find container \"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\": container with ID starting with a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.504747 5025 scope.go:117] "RemoveContainer" containerID="8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.504990 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4"} err="failed to get container status \"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\": rpc error: code = NotFound desc = could not find container \"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\": container with ID starting with 8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.505029 5025 scope.go:117] "RemoveContainer" containerID="7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.505463 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906"} err="failed to get container status \"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\": rpc error: code = NotFound desc = could not find container \"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\": container with ID starting with 7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.505484 5025 scope.go:117] "RemoveContainer" containerID="b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.505705 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95"} err="failed to get container status \"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\": rpc error: code = NotFound desc = could not find container \"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\": container with ID starting with b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.505737 5025 scope.go:117] "RemoveContainer" containerID="9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.506084 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae"} err="failed to get container status \"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\": rpc error: code = NotFound desc = could not find container \"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\": container with ID starting with 9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.506104 5025 scope.go:117] "RemoveContainer" containerID="e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.506353 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495"} err="failed to get container status \"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\": rpc error: code = NotFound desc = could not find container \"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\": container with ID starting with e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.506379 5025 scope.go:117] "RemoveContainer" containerID="01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.506611 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a"} err="failed to get container status \"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a\": rpc error: code = NotFound desc = could not find container \"01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a\": container with ID starting with 01361f4372393ddcb7dc392d276f049d8921d121d0eb41ed933dfd704acc418a not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.506635 5025 scope.go:117] "RemoveContainer" containerID="a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.506872 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2"} err="failed to get container status \"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\": rpc error: code = NotFound desc = could not find container \"a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2\": container with ID starting with a8ebcf81b9a4657fdb4866a48e9027c6568eabba140cc3075e223843698772d2 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.506894 5025 scope.go:117] "RemoveContainer" containerID="80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.507104 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9"} err="failed to get container status \"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\": rpc error: code = NotFound desc = could not find container \"80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9\": container with ID starting with 80bb3c23ba0b3ca6efa67d3bea75ebb06f9099b38cda9851a3991a5bac289ab9 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.507127 5025 scope.go:117] "RemoveContainer" containerID="a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.507493 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4"} err="failed to get container status \"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\": rpc error: code = NotFound desc = could not find container \"a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4\": container with ID starting with a1f0e04167bcf10b59dd25a2711f65bfbdbbb7c0ca55e722f4ff41f20023f8a4 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.507516 5025 scope.go:117] "RemoveContainer" containerID="a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.507727 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648"} err="failed to get container status \"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\": rpc error: code = NotFound desc = could not find container \"a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648\": container with ID starting with a4646bc038e21fbc58bb458f3dd3a153c6f23feb953fe427f1be34001869d648 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.507751 5025 scope.go:117] "RemoveContainer" containerID="8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.508225 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4"} err="failed to get container status \"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\": rpc error: code = NotFound desc = could not find container \"8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4\": container with ID starting with 8f01de113a9e09bf0a9cf1f4352bcf715fd1553c39b4ca6ca581433c430ba3f4 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.508251 5025 scope.go:117] "RemoveContainer" containerID="7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.508513 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906"} err="failed to get container status \"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\": rpc error: code = NotFound desc = could not find container \"7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906\": container with ID starting with 7f81bda7777089edc468e6b6a9be872d1fd438831bc9918fd5c77f93cb970906 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.508542 5025 scope.go:117] "RemoveContainer" containerID="b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.508731 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95"} err="failed to get container status \"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\": rpc error: code = NotFound desc = could not find container \"b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95\": container with ID starting with b6f36c4ea9b29e18a10950855fb6d55606f33e44fea8027870ded49209575a95 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.508757 5025 scope.go:117] "RemoveContainer" containerID="9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.508958 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae"} err="failed to get container status \"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\": rpc error: code = NotFound desc = could not find container \"9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae\": container with ID starting with 9bb12a6c1103283f77f4b877c06cd9bba3852ac905e12a8125e28265da2228ae not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.508988 5025 scope.go:117] "RemoveContainer" containerID="e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.509200 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495"} err="failed to get container status \"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\": rpc error: code = NotFound desc = could not find container \"e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495\": container with ID starting with e22dad8010191102c937c26488c3a13fcbfcf7d1304f5a33ed2028ae1f745495 not found: ID does not exist" Oct 04 10:45:07 crc kubenswrapper[5025]: I1004 10:45:07.536764 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:07 crc kubenswrapper[5025]: W1004 10:45:07.552048 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12c41f75_e5d5_47e3_a4bb_162c922a5912.slice/crio-833c25e0393f8bc1f78cf7d13f365620e49c1dbe81f6b2c8a5985553f0590f44 WatchSource:0}: Error finding container 833c25e0393f8bc1f78cf7d13f365620e49c1dbe81f6b2c8a5985553f0590f44: Status 404 returned error can't find the container with id 833c25e0393f8bc1f78cf7d13f365620e49c1dbe81f6b2c8a5985553f0590f44 Oct 04 10:45:08 crc kubenswrapper[5025]: I1004 10:45:08.349474 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n49fv_8d214492-7532-4f72-b032-b3cc1ae19473/kube-multus/2.log" Oct 04 10:45:08 crc kubenswrapper[5025]: I1004 10:45:08.352054 5025 generic.go:334] "Generic (PLEG): container finished" podID="12c41f75-e5d5-47e3-a4bb-162c922a5912" containerID="e43fe905f2216f2f30a0ad60c3418b96d1656cb58344249ad4054df7b23069d6" exitCode=0 Oct 04 10:45:08 crc kubenswrapper[5025]: I1004 10:45:08.352130 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" event={"ID":"12c41f75-e5d5-47e3-a4bb-162c922a5912","Type":"ContainerDied","Data":"e43fe905f2216f2f30a0ad60c3418b96d1656cb58344249ad4054df7b23069d6"} Oct 04 10:45:08 crc kubenswrapper[5025]: I1004 10:45:08.352174 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" event={"ID":"12c41f75-e5d5-47e3-a4bb-162c922a5912","Type":"ContainerStarted","Data":"833c25e0393f8bc1f78cf7d13f365620e49c1dbe81f6b2c8a5985553f0590f44"} Oct 04 10:45:08 crc kubenswrapper[5025]: I1004 10:45:08.418818 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="607a1d66-62fc-4dba-9c44-6798f087fb5c" path="/var/lib/kubelet/pods/607a1d66-62fc-4dba-9c44-6798f087fb5c/volumes" Oct 04 10:45:09 crc kubenswrapper[5025]: I1004 10:45:09.360906 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" event={"ID":"12c41f75-e5d5-47e3-a4bb-162c922a5912","Type":"ContainerStarted","Data":"d5d622eeb027cae3a512ef6314b190310e7a0c7c4c9644f15640295c55530f06"} Oct 04 10:45:09 crc kubenswrapper[5025]: I1004 10:45:09.361738 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" event={"ID":"12c41f75-e5d5-47e3-a4bb-162c922a5912","Type":"ContainerStarted","Data":"71c90794d5cfa5fe3e25f19fc87733796884b4e5a91c1facbee51204ad36ca0d"} Oct 04 10:45:09 crc kubenswrapper[5025]: I1004 10:45:09.361825 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" event={"ID":"12c41f75-e5d5-47e3-a4bb-162c922a5912","Type":"ContainerStarted","Data":"ee11a126bb4ebe80dcf2e5aa78f1fb1af84d0d3dd231578753d9f1a4f0172aac"} Oct 04 10:45:09 crc kubenswrapper[5025]: I1004 10:45:09.361898 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" event={"ID":"12c41f75-e5d5-47e3-a4bb-162c922a5912","Type":"ContainerStarted","Data":"7cb9d76e445a4588e9a224cce3316b5f5a39eb0f99f51515d2a78d2f3bdb009d"} Oct 04 10:45:09 crc kubenswrapper[5025]: I1004 10:45:09.361972 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" event={"ID":"12c41f75-e5d5-47e3-a4bb-162c922a5912","Type":"ContainerStarted","Data":"c0ac24f7982b1dab2adf7acbe186b658d436ab7ba73caa8cd3d7f673f2a3c461"} Oct 04 10:45:09 crc kubenswrapper[5025]: I1004 10:45:09.362077 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" event={"ID":"12c41f75-e5d5-47e3-a4bb-162c922a5912","Type":"ContainerStarted","Data":"0f4f33725a61bd72a4f65fb3250afb159cfc6b091026ba17bbeaf5cd89fff833"} Oct 04 10:45:12 crc kubenswrapper[5025]: I1004 10:45:12.390044 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" event={"ID":"12c41f75-e5d5-47e3-a4bb-162c922a5912","Type":"ContainerStarted","Data":"0faeaf3253d3e1dc1e1c4d74a8f13d3fb354cbc7d3d376def97a5dba58426ecc"} Oct 04 10:45:14 crc kubenswrapper[5025]: I1004 10:45:14.406238 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" event={"ID":"12c41f75-e5d5-47e3-a4bb-162c922a5912","Type":"ContainerStarted","Data":"0c0e3f071a73b37d9fe4c8d8a2165977f99a91dcf0f1f10c1d18459ee49a773c"} Oct 04 10:45:14 crc kubenswrapper[5025]: I1004 10:45:14.406618 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:14 crc kubenswrapper[5025]: I1004 10:45:14.406654 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:14 crc kubenswrapper[5025]: I1004 10:45:14.406664 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:14 crc kubenswrapper[5025]: I1004 10:45:14.443226 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:14 crc kubenswrapper[5025]: I1004 10:45:14.443980 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" podStartSLOduration=7.44396905 podStartE2EDuration="7.44396905s" podCreationTimestamp="2025-10-04 10:45:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:45:14.440323742 +0000 UTC m=+642.865290622" watchObservedRunningTime="2025-10-04 10:45:14.44396905 +0000 UTC m=+642.868935930" Oct 04 10:45:14 crc kubenswrapper[5025]: I1004 10:45:14.447665 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:18 crc kubenswrapper[5025]: I1004 10:45:18.411171 5025 scope.go:117] "RemoveContainer" containerID="f1d8ee42e06786ad69717a96ff5cdc23094d6c2bb0f8a56b8a35ed098ba52011" Oct 04 10:45:18 crc kubenswrapper[5025]: E1004 10:45:18.411775 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-n49fv_openshift-multus(8d214492-7532-4f72-b032-b3cc1ae19473)\"" pod="openshift-multus/multus-n49fv" podUID="8d214492-7532-4f72-b032-b3cc1ae19473" Oct 04 10:45:32 crc kubenswrapper[5025]: I1004 10:45:32.416260 5025 scope.go:117] "RemoveContainer" containerID="f1d8ee42e06786ad69717a96ff5cdc23094d6c2bb0f8a56b8a35ed098ba52011" Oct 04 10:45:33 crc kubenswrapper[5025]: I1004 10:45:33.526415 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-n49fv_8d214492-7532-4f72-b032-b3cc1ae19473/kube-multus/2.log" Oct 04 10:45:33 crc kubenswrapper[5025]: I1004 10:45:33.526973 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-n49fv" event={"ID":"8d214492-7532-4f72-b032-b3cc1ae19473","Type":"ContainerStarted","Data":"4b6ef36604817b450c796be6118d7ffea196a1340e894332fa56a8d581eb6496"} Oct 04 10:45:37 crc kubenswrapper[5025]: I1004 10:45:37.564827 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zmq6r" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.055671 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr"] Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.057136 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.059191 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.066893 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr"] Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.160999 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.161652 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.161765 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d88n7\" (UniqueName: \"kubernetes.io/projected/33dbe626-f29f-4939-bcdc-c04109e9820c-kube-api-access-d88n7\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.263100 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d88n7\" (UniqueName: \"kubernetes.io/projected/33dbe626-f29f-4939-bcdc-c04109e9820c-kube-api-access-d88n7\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.263327 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.263350 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.263794 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.264104 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.282181 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d88n7\" (UniqueName: \"kubernetes.io/projected/33dbe626-f29f-4939-bcdc-c04109e9820c-kube-api-access-d88n7\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.377193 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:46 crc kubenswrapper[5025]: I1004 10:45:46.619056 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr"] Oct 04 10:45:46 crc kubenswrapper[5025]: W1004 10:45:46.620711 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33dbe626_f29f_4939_bcdc_c04109e9820c.slice/crio-a2f25ea771fb1c5270f08dae68649afd532063992bd993d65600eef00961aeb1 WatchSource:0}: Error finding container a2f25ea771fb1c5270f08dae68649afd532063992bd993d65600eef00961aeb1: Status 404 returned error can't find the container with id a2f25ea771fb1c5270f08dae68649afd532063992bd993d65600eef00961aeb1 Oct 04 10:45:47 crc kubenswrapper[5025]: I1004 10:45:47.607941 5025 generic.go:334] "Generic (PLEG): container finished" podID="33dbe626-f29f-4939-bcdc-c04109e9820c" containerID="6dbfac70c691b060721abe0e59c1fb511be6914f8d1fdfcaa6472189eaac3bfd" exitCode=0 Oct 04 10:45:47 crc kubenswrapper[5025]: I1004 10:45:47.608040 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" event={"ID":"33dbe626-f29f-4939-bcdc-c04109e9820c","Type":"ContainerDied","Data":"6dbfac70c691b060721abe0e59c1fb511be6914f8d1fdfcaa6472189eaac3bfd"} Oct 04 10:45:47 crc kubenswrapper[5025]: I1004 10:45:47.608359 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" event={"ID":"33dbe626-f29f-4939-bcdc-c04109e9820c","Type":"ContainerStarted","Data":"a2f25ea771fb1c5270f08dae68649afd532063992bd993d65600eef00961aeb1"} Oct 04 10:45:50 crc kubenswrapper[5025]: I1004 10:45:50.629692 5025 generic.go:334] "Generic (PLEG): container finished" podID="33dbe626-f29f-4939-bcdc-c04109e9820c" containerID="b70657a184055ac37af3ef704195a4ff923e62c185fb245729c3c567b622d35b" exitCode=0 Oct 04 10:45:50 crc kubenswrapper[5025]: I1004 10:45:50.629798 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" event={"ID":"33dbe626-f29f-4939-bcdc-c04109e9820c","Type":"ContainerDied","Data":"b70657a184055ac37af3ef704195a4ff923e62c185fb245729c3c567b622d35b"} Oct 04 10:45:51 crc kubenswrapper[5025]: E1004 10:45:51.093203 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33dbe626_f29f_4939_bcdc_c04109e9820c.slice/crio-ffaae5eae78d7ad385ff6cb32704bbfa23330856e1b189aa6ffe757c20205ed0.scope\": RecentStats: unable to find data in memory cache]" Oct 04 10:45:51 crc kubenswrapper[5025]: I1004 10:45:51.637117 5025 generic.go:334] "Generic (PLEG): container finished" podID="33dbe626-f29f-4939-bcdc-c04109e9820c" containerID="ffaae5eae78d7ad385ff6cb32704bbfa23330856e1b189aa6ffe757c20205ed0" exitCode=0 Oct 04 10:45:51 crc kubenswrapper[5025]: I1004 10:45:51.637172 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" event={"ID":"33dbe626-f29f-4939-bcdc-c04109e9820c","Type":"ContainerDied","Data":"ffaae5eae78d7ad385ff6cb32704bbfa23330856e1b189aa6ffe757c20205ed0"} Oct 04 10:45:52 crc kubenswrapper[5025]: I1004 10:45:52.930001 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.089873 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d88n7\" (UniqueName: \"kubernetes.io/projected/33dbe626-f29f-4939-bcdc-c04109e9820c-kube-api-access-d88n7\") pod \"33dbe626-f29f-4939-bcdc-c04109e9820c\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.089999 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-util\") pod \"33dbe626-f29f-4939-bcdc-c04109e9820c\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.090131 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-bundle\") pod \"33dbe626-f29f-4939-bcdc-c04109e9820c\" (UID: \"33dbe626-f29f-4939-bcdc-c04109e9820c\") " Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.090541 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-bundle" (OuterVolumeSpecName: "bundle") pod "33dbe626-f29f-4939-bcdc-c04109e9820c" (UID: "33dbe626-f29f-4939-bcdc-c04109e9820c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.101083 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-util" (OuterVolumeSpecName: "util") pod "33dbe626-f29f-4939-bcdc-c04109e9820c" (UID: "33dbe626-f29f-4939-bcdc-c04109e9820c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.101952 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33dbe626-f29f-4939-bcdc-c04109e9820c-kube-api-access-d88n7" (OuterVolumeSpecName: "kube-api-access-d88n7") pod "33dbe626-f29f-4939-bcdc-c04109e9820c" (UID: "33dbe626-f29f-4939-bcdc-c04109e9820c"). InnerVolumeSpecName "kube-api-access-d88n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.192163 5025 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.192201 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d88n7\" (UniqueName: \"kubernetes.io/projected/33dbe626-f29f-4939-bcdc-c04109e9820c-kube-api-access-d88n7\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.192216 5025 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/33dbe626-f29f-4939-bcdc-c04109e9820c-util\") on node \"crc\" DevicePath \"\"" Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.653825 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" event={"ID":"33dbe626-f29f-4939-bcdc-c04109e9820c","Type":"ContainerDied","Data":"a2f25ea771fb1c5270f08dae68649afd532063992bd993d65600eef00961aeb1"} Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.653874 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2f25ea771fb1c5270f08dae68649afd532063992bd993d65600eef00961aeb1" Oct 04 10:45:53 crc kubenswrapper[5025]: I1004 10:45:53.653958 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr" Oct 04 10:45:54 crc kubenswrapper[5025]: I1004 10:45:54.957404 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6"] Oct 04 10:45:54 crc kubenswrapper[5025]: E1004 10:45:54.957936 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33dbe626-f29f-4939-bcdc-c04109e9820c" containerName="util" Oct 04 10:45:54 crc kubenswrapper[5025]: I1004 10:45:54.957951 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="33dbe626-f29f-4939-bcdc-c04109e9820c" containerName="util" Oct 04 10:45:54 crc kubenswrapper[5025]: E1004 10:45:54.957969 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33dbe626-f29f-4939-bcdc-c04109e9820c" containerName="extract" Oct 04 10:45:54 crc kubenswrapper[5025]: I1004 10:45:54.957977 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="33dbe626-f29f-4939-bcdc-c04109e9820c" containerName="extract" Oct 04 10:45:54 crc kubenswrapper[5025]: E1004 10:45:54.957990 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33dbe626-f29f-4939-bcdc-c04109e9820c" containerName="pull" Oct 04 10:45:54 crc kubenswrapper[5025]: I1004 10:45:54.957998 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="33dbe626-f29f-4939-bcdc-c04109e9820c" containerName="pull" Oct 04 10:45:54 crc kubenswrapper[5025]: I1004 10:45:54.958128 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="33dbe626-f29f-4939-bcdc-c04109e9820c" containerName="extract" Oct 04 10:45:54 crc kubenswrapper[5025]: I1004 10:45:54.958608 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6" Oct 04 10:45:54 crc kubenswrapper[5025]: I1004 10:45:54.960356 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 04 10:45:54 crc kubenswrapper[5025]: I1004 10:45:54.963980 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-4ttq9" Oct 04 10:45:54 crc kubenswrapper[5025]: I1004 10:45:54.964317 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 04 10:45:54 crc kubenswrapper[5025]: I1004 10:45:54.969143 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6"] Oct 04 10:45:55 crc kubenswrapper[5025]: I1004 10:45:55.114682 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjrbg\" (UniqueName: \"kubernetes.io/projected/3b47ddfd-e0dd-4a6f-9879-505058ddc2b3-kube-api-access-jjrbg\") pod \"nmstate-operator-858ddd8f98-dvdx6\" (UID: \"3b47ddfd-e0dd-4a6f-9879-505058ddc2b3\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6" Oct 04 10:45:55 crc kubenswrapper[5025]: I1004 10:45:55.216276 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjrbg\" (UniqueName: \"kubernetes.io/projected/3b47ddfd-e0dd-4a6f-9879-505058ddc2b3-kube-api-access-jjrbg\") pod \"nmstate-operator-858ddd8f98-dvdx6\" (UID: \"3b47ddfd-e0dd-4a6f-9879-505058ddc2b3\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6" Oct 04 10:45:55 crc kubenswrapper[5025]: I1004 10:45:55.244667 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjrbg\" (UniqueName: \"kubernetes.io/projected/3b47ddfd-e0dd-4a6f-9879-505058ddc2b3-kube-api-access-jjrbg\") pod \"nmstate-operator-858ddd8f98-dvdx6\" (UID: \"3b47ddfd-e0dd-4a6f-9879-505058ddc2b3\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6" Oct 04 10:45:55 crc kubenswrapper[5025]: I1004 10:45:55.272610 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6" Oct 04 10:45:55 crc kubenswrapper[5025]: I1004 10:45:55.453391 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6"] Oct 04 10:45:55 crc kubenswrapper[5025]: I1004 10:45:55.664926 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6" event={"ID":"3b47ddfd-e0dd-4a6f-9879-505058ddc2b3","Type":"ContainerStarted","Data":"2a574b09d8ed11559ca4d2b07aaa8d1a331963b874d93059d2d76ef4256e46e7"} Oct 04 10:45:58 crc kubenswrapper[5025]: I1004 10:45:58.693081 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6" event={"ID":"3b47ddfd-e0dd-4a6f-9879-505058ddc2b3","Type":"ContainerStarted","Data":"2cc5695baa25fc1d2c6d7d43738847f2d083cd78462f8dd95820d08b358abd39"} Oct 04 10:45:58 crc kubenswrapper[5025]: I1004 10:45:58.713475 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-dvdx6" podStartSLOduration=2.606864234 podStartE2EDuration="4.713430237s" podCreationTimestamp="2025-10-04 10:45:54 +0000 UTC" firstStartedPulling="2025-10-04 10:45:55.461547376 +0000 UTC m=+683.886514256" lastFinishedPulling="2025-10-04 10:45:57.568113379 +0000 UTC m=+685.993080259" observedRunningTime="2025-10-04 10:45:58.708241804 +0000 UTC m=+687.133208684" watchObservedRunningTime="2025-10-04 10:45:58.713430237 +0000 UTC m=+687.138397127" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.793997 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr"] Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.795361 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.796803 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xlkkr" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.802874 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs"] Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.803739 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.810365 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.813722 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr"] Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.829817 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs"] Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.855878 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-ksclx"] Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.856694 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.875003 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75qvm\" (UniqueName: \"kubernetes.io/projected/4f852867-9d9b-4255-898c-3b26aabf3243-kube-api-access-75qvm\") pod \"nmstate-webhook-6cdbc54649-g2ggs\" (UID: \"4f852867-9d9b-4255-898c-3b26aabf3243\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.875608 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4f852867-9d9b-4255-898c-3b26aabf3243-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-g2ggs\" (UID: \"4f852867-9d9b-4255-898c-3b26aabf3243\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.876378 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7nzm\" (UniqueName: \"kubernetes.io/projected/7604348c-0bc1-4932-8ff6-00b3098c7326-kube-api-access-r7nzm\") pod \"nmstate-metrics-fdff9cb8d-55hjr\" (UID: \"7604348c-0bc1-4932-8ff6-00b3098c7326\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.945265 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64"] Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.946260 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.948775 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.949142 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.949351 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-qkcz9" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.966341 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64"] Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.981347 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1c998df0-be41-4019-bba6-5e1b4f58a10e-nmstate-lock\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.981394 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mrdm\" (UniqueName: \"kubernetes.io/projected/1c998df0-be41-4019-bba6-5e1b4f58a10e-kube-api-access-2mrdm\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.981460 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1c998df0-be41-4019-bba6-5e1b4f58a10e-ovs-socket\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.981499 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75qvm\" (UniqueName: \"kubernetes.io/projected/4f852867-9d9b-4255-898c-3b26aabf3243-kube-api-access-75qvm\") pod \"nmstate-webhook-6cdbc54649-g2ggs\" (UID: \"4f852867-9d9b-4255-898c-3b26aabf3243\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.981547 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4f852867-9d9b-4255-898c-3b26aabf3243-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-g2ggs\" (UID: \"4f852867-9d9b-4255-898c-3b26aabf3243\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.981574 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7nzm\" (UniqueName: \"kubernetes.io/projected/7604348c-0bc1-4932-8ff6-00b3098c7326-kube-api-access-r7nzm\") pod \"nmstate-metrics-fdff9cb8d-55hjr\" (UID: \"7604348c-0bc1-4932-8ff6-00b3098c7326\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr" Oct 04 10:45:59 crc kubenswrapper[5025]: I1004 10:45:59.981598 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1c998df0-be41-4019-bba6-5e1b4f58a10e-dbus-socket\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:45:59 crc kubenswrapper[5025]: E1004 10:45:59.981742 5025 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 04 10:45:59 crc kubenswrapper[5025]: E1004 10:45:59.981796 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4f852867-9d9b-4255-898c-3b26aabf3243-tls-key-pair podName:4f852867-9d9b-4255-898c-3b26aabf3243 nodeName:}" failed. No retries permitted until 2025-10-04 10:46:00.481775815 +0000 UTC m=+688.906742695 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/4f852867-9d9b-4255-898c-3b26aabf3243-tls-key-pair") pod "nmstate-webhook-6cdbc54649-g2ggs" (UID: "4f852867-9d9b-4255-898c-3b26aabf3243") : secret "openshift-nmstate-webhook" not found Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.001769 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75qvm\" (UniqueName: \"kubernetes.io/projected/4f852867-9d9b-4255-898c-3b26aabf3243-kube-api-access-75qvm\") pod \"nmstate-webhook-6cdbc54649-g2ggs\" (UID: \"4f852867-9d9b-4255-898c-3b26aabf3243\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.003661 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7nzm\" (UniqueName: \"kubernetes.io/projected/7604348c-0bc1-4932-8ff6-00b3098c7326-kube-api-access-r7nzm\") pod \"nmstate-metrics-fdff9cb8d-55hjr\" (UID: \"7604348c-0bc1-4932-8ff6-00b3098c7326\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.083169 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/56180570-2df4-4eee-93dd-d403e7c4a24f-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-mtz64\" (UID: \"56180570-2df4-4eee-93dd-d403e7c4a24f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.083244 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1c998df0-be41-4019-bba6-5e1b4f58a10e-dbus-socket\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.083267 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1c998df0-be41-4019-bba6-5e1b4f58a10e-nmstate-lock\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.083289 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mrdm\" (UniqueName: \"kubernetes.io/projected/1c998df0-be41-4019-bba6-5e1b4f58a10e-kube-api-access-2mrdm\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.083313 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk4zq\" (UniqueName: \"kubernetes.io/projected/56180570-2df4-4eee-93dd-d403e7c4a24f-kube-api-access-mk4zq\") pod \"nmstate-console-plugin-6b874cbd85-mtz64\" (UID: \"56180570-2df4-4eee-93dd-d403e7c4a24f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.083354 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1c998df0-be41-4019-bba6-5e1b4f58a10e-ovs-socket\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.083371 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/56180570-2df4-4eee-93dd-d403e7c4a24f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-mtz64\" (UID: \"56180570-2df4-4eee-93dd-d403e7c4a24f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.083699 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/1c998df0-be41-4019-bba6-5e1b4f58a10e-ovs-socket\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.083719 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/1c998df0-be41-4019-bba6-5e1b4f58a10e-nmstate-lock\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.083736 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/1c998df0-be41-4019-bba6-5e1b4f58a10e-dbus-socket\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.108765 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mrdm\" (UniqueName: \"kubernetes.io/projected/1c998df0-be41-4019-bba6-5e1b4f58a10e-kube-api-access-2mrdm\") pod \"nmstate-handler-ksclx\" (UID: \"1c998df0-be41-4019-bba6-5e1b4f58a10e\") " pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.117609 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.127288 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-76b6b6b6bb-n2cbx"] Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.128862 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.142212 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76b6b6b6bb-n2cbx"] Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.183080 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.184286 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/56180570-2df4-4eee-93dd-d403e7c4a24f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-mtz64\" (UID: \"56180570-2df4-4eee-93dd-d403e7c4a24f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.184969 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/56180570-2df4-4eee-93dd-d403e7c4a24f-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-mtz64\" (UID: \"56180570-2df4-4eee-93dd-d403e7c4a24f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.185136 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk4zq\" (UniqueName: \"kubernetes.io/projected/56180570-2df4-4eee-93dd-d403e7c4a24f-kube-api-access-mk4zq\") pod \"nmstate-console-plugin-6b874cbd85-mtz64\" (UID: \"56180570-2df4-4eee-93dd-d403e7c4a24f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.185992 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/56180570-2df4-4eee-93dd-d403e7c4a24f-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-mtz64\" (UID: \"56180570-2df4-4eee-93dd-d403e7c4a24f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.188349 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/56180570-2df4-4eee-93dd-d403e7c4a24f-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-mtz64\" (UID: \"56180570-2df4-4eee-93dd-d403e7c4a24f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.200734 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk4zq\" (UniqueName: \"kubernetes.io/projected/56180570-2df4-4eee-93dd-d403e7c4a24f-kube-api-access-mk4zq\") pod \"nmstate-console-plugin-6b874cbd85-mtz64\" (UID: \"56180570-2df4-4eee-93dd-d403e7c4a24f\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.266750 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.286218 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-service-ca\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.286291 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-trusted-ca-bundle\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.286399 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03716ccb-771d-4f62-b169-10657378d569-console-serving-cert\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.286504 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-oauth-serving-cert\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.286543 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btmtf\" (UniqueName: \"kubernetes.io/projected/03716ccb-771d-4f62-b169-10657378d569-kube-api-access-btmtf\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.286574 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03716ccb-771d-4f62-b169-10657378d569-console-oauth-config\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.286591 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-console-config\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.328847 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr"] Oct 04 10:46:00 crc kubenswrapper[5025]: W1004 10:46:00.339555 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7604348c_0bc1_4932_8ff6_00b3098c7326.slice/crio-8ac977d5b25cb9501410916d77ff8706b0b2e683aa003ac2964b22e6bbb6dcef WatchSource:0}: Error finding container 8ac977d5b25cb9501410916d77ff8706b0b2e683aa003ac2964b22e6bbb6dcef: Status 404 returned error can't find the container with id 8ac977d5b25cb9501410916d77ff8706b0b2e683aa003ac2964b22e6bbb6dcef Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.388058 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-service-ca\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.388118 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-trusted-ca-bundle\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.388152 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03716ccb-771d-4f62-b169-10657378d569-console-serving-cert\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.388211 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-oauth-serving-cert\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.388242 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btmtf\" (UniqueName: \"kubernetes.io/projected/03716ccb-771d-4f62-b169-10657378d569-kube-api-access-btmtf\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.388271 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03716ccb-771d-4f62-b169-10657378d569-console-oauth-config\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.388293 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-console-config\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.389436 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-service-ca\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.389448 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-oauth-serving-cert\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.389467 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-console-config\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.389548 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/03716ccb-771d-4f62-b169-10657378d569-trusted-ca-bundle\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.393548 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/03716ccb-771d-4f62-b169-10657378d569-console-serving-cert\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.399385 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/03716ccb-771d-4f62-b169-10657378d569-console-oauth-config\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.407898 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btmtf\" (UniqueName: \"kubernetes.io/projected/03716ccb-771d-4f62-b169-10657378d569-kube-api-access-btmtf\") pod \"console-76b6b6b6bb-n2cbx\" (UID: \"03716ccb-771d-4f62-b169-10657378d569\") " pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.463773 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64"] Oct 04 10:46:00 crc kubenswrapper[5025]: W1004 10:46:00.468264 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56180570_2df4_4eee_93dd_d403e7c4a24f.slice/crio-35d592ecd8979d2d4a5a86a65483f8290cb6e4f189b6c286785e2ba2094e0a3c WatchSource:0}: Error finding container 35d592ecd8979d2d4a5a86a65483f8290cb6e4f189b6c286785e2ba2094e0a3c: Status 404 returned error can't find the container with id 35d592ecd8979d2d4a5a86a65483f8290cb6e4f189b6c286785e2ba2094e0a3c Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.489000 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.489481 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4f852867-9d9b-4255-898c-3b26aabf3243-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-g2ggs\" (UID: \"4f852867-9d9b-4255-898c-3b26aabf3243\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.492592 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4f852867-9d9b-4255-898c-3b26aabf3243-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-g2ggs\" (UID: \"4f852867-9d9b-4255-898c-3b26aabf3243\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.670413 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-76b6b6b6bb-n2cbx"] Oct 04 10:46:00 crc kubenswrapper[5025]: W1004 10:46:00.681591 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03716ccb_771d_4f62_b169_10657378d569.slice/crio-e325e4e621e806d06ba2deb91deef86ce3e571d421ff2bc933a4cd01c32145a1 WatchSource:0}: Error finding container e325e4e621e806d06ba2deb91deef86ce3e571d421ff2bc933a4cd01c32145a1: Status 404 returned error can't find the container with id e325e4e621e806d06ba2deb91deef86ce3e571d421ff2bc933a4cd01c32145a1 Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.704389 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-ksclx" event={"ID":"1c998df0-be41-4019-bba6-5e1b4f58a10e","Type":"ContainerStarted","Data":"82f84216dad16fb040a642e10ee5f5c5dd7bd834ecb284e9e8f968c51201e53d"} Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.705912 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" event={"ID":"56180570-2df4-4eee-93dd-d403e7c4a24f","Type":"ContainerStarted","Data":"35d592ecd8979d2d4a5a86a65483f8290cb6e4f189b6c286785e2ba2094e0a3c"} Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.707031 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr" event={"ID":"7604348c-0bc1-4932-8ff6-00b3098c7326","Type":"ContainerStarted","Data":"8ac977d5b25cb9501410916d77ff8706b0b2e683aa003ac2964b22e6bbb6dcef"} Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.707935 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76b6b6b6bb-n2cbx" event={"ID":"03716ccb-771d-4f62-b169-10657378d569","Type":"ContainerStarted","Data":"e325e4e621e806d06ba2deb91deef86ce3e571d421ff2bc933a4cd01c32145a1"} Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.732998 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:46:00 crc kubenswrapper[5025]: I1004 10:46:00.945882 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs"] Oct 04 10:46:01 crc kubenswrapper[5025]: I1004 10:46:01.721766 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" event={"ID":"4f852867-9d9b-4255-898c-3b26aabf3243","Type":"ContainerStarted","Data":"8175fcd073ed8ab550a22364e6649c8b5606bc6da68e7ffd700928d3e060fe99"} Oct 04 10:46:01 crc kubenswrapper[5025]: I1004 10:46:01.724221 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-76b6b6b6bb-n2cbx" event={"ID":"03716ccb-771d-4f62-b169-10657378d569","Type":"ContainerStarted","Data":"29150be321b77b747f35a1e167fc1551eb74bb967d64cef7c85a3771e42e26d4"} Oct 04 10:46:01 crc kubenswrapper[5025]: I1004 10:46:01.747487 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-76b6b6b6bb-n2cbx" podStartSLOduration=1.74746326 podStartE2EDuration="1.74746326s" podCreationTimestamp="2025-10-04 10:46:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:46:01.741547826 +0000 UTC m=+690.166514716" watchObservedRunningTime="2025-10-04 10:46:01.74746326 +0000 UTC m=+690.172430140" Oct 04 10:46:03 crc kubenswrapper[5025]: I1004 10:46:03.738125 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-ksclx" event={"ID":"1c998df0-be41-4019-bba6-5e1b4f58a10e","Type":"ContainerStarted","Data":"068549fcb090b5bbcafc73cb389728bbc883334431e8094f786654b92eca6525"} Oct 04 10:46:03 crc kubenswrapper[5025]: I1004 10:46:03.738611 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:03 crc kubenswrapper[5025]: I1004 10:46:03.740362 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" event={"ID":"56180570-2df4-4eee-93dd-d403e7c4a24f","Type":"ContainerStarted","Data":"845f5e7fae5161941716c6e3547c02715fdaef463e421874e42c50f20bb64d61"} Oct 04 10:46:03 crc kubenswrapper[5025]: I1004 10:46:03.741757 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr" event={"ID":"7604348c-0bc1-4932-8ff6-00b3098c7326","Type":"ContainerStarted","Data":"61e11172c1f6d2c70d26032c230f5e88f379926a7dbac7579b440e2d41ccf7fc"} Oct 04 10:46:03 crc kubenswrapper[5025]: I1004 10:46:03.743216 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" event={"ID":"4f852867-9d9b-4255-898c-3b26aabf3243","Type":"ContainerStarted","Data":"2fb1ac3bedc6c7a45226a45b812446f8bbd39ad366ee6e1667c3828f3cbdc73c"} Oct 04 10:46:03 crc kubenswrapper[5025]: I1004 10:46:03.743369 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:46:03 crc kubenswrapper[5025]: I1004 10:46:03.758534 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-ksclx" podStartSLOduration=1.7564658419999999 podStartE2EDuration="4.758515984s" podCreationTimestamp="2025-10-04 10:45:59 +0000 UTC" firstStartedPulling="2025-10-04 10:46:00.204601419 +0000 UTC m=+688.629568299" lastFinishedPulling="2025-10-04 10:46:03.206651521 +0000 UTC m=+691.631618441" observedRunningTime="2025-10-04 10:46:03.756546596 +0000 UTC m=+692.181513486" watchObservedRunningTime="2025-10-04 10:46:03.758515984 +0000 UTC m=+692.183482864" Oct 04 10:46:03 crc kubenswrapper[5025]: I1004 10:46:03.771526 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" podStartSLOduration=2.488127733 podStartE2EDuration="4.771506846s" podCreationTimestamp="2025-10-04 10:45:59 +0000 UTC" firstStartedPulling="2025-10-04 10:46:00.955765404 +0000 UTC m=+689.380732274" lastFinishedPulling="2025-10-04 10:46:03.239144477 +0000 UTC m=+691.664111387" observedRunningTime="2025-10-04 10:46:03.770668752 +0000 UTC m=+692.195635632" watchObservedRunningTime="2025-10-04 10:46:03.771506846 +0000 UTC m=+692.196473726" Oct 04 10:46:03 crc kubenswrapper[5025]: I1004 10:46:03.789914 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-mtz64" podStartSLOduration=2.059550777 podStartE2EDuration="4.789893227s" podCreationTimestamp="2025-10-04 10:45:59 +0000 UTC" firstStartedPulling="2025-10-04 10:46:00.471436538 +0000 UTC m=+688.896403418" lastFinishedPulling="2025-10-04 10:46:03.201778988 +0000 UTC m=+691.626745868" observedRunningTime="2025-10-04 10:46:03.789522676 +0000 UTC m=+692.214489566" watchObservedRunningTime="2025-10-04 10:46:03.789893227 +0000 UTC m=+692.214860107" Oct 04 10:46:05 crc kubenswrapper[5025]: I1004 10:46:05.756707 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr" event={"ID":"7604348c-0bc1-4932-8ff6-00b3098c7326","Type":"ContainerStarted","Data":"80dd00c1b3340d9e1724eaf13728e7ed1dc0eed0ef329f66ca9b7e68ac6841d6"} Oct 04 10:46:05 crc kubenswrapper[5025]: I1004 10:46:05.790770 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-55hjr" podStartSLOduration=1.697540569 podStartE2EDuration="6.79073642s" podCreationTimestamp="2025-10-04 10:45:59 +0000 UTC" firstStartedPulling="2025-10-04 10:46:00.34265214 +0000 UTC m=+688.767619020" lastFinishedPulling="2025-10-04 10:46:05.435847991 +0000 UTC m=+693.860814871" observedRunningTime="2025-10-04 10:46:05.780614292 +0000 UTC m=+694.205581182" watchObservedRunningTime="2025-10-04 10:46:05.79073642 +0000 UTC m=+694.215703340" Oct 04 10:46:10 crc kubenswrapper[5025]: I1004 10:46:10.214523 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-ksclx" Oct 04 10:46:10 crc kubenswrapper[5025]: I1004 10:46:10.489265 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:10 crc kubenswrapper[5025]: I1004 10:46:10.489682 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:10 crc kubenswrapper[5025]: I1004 10:46:10.494385 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:10 crc kubenswrapper[5025]: I1004 10:46:10.799065 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-76b6b6b6bb-n2cbx" Oct 04 10:46:10 crc kubenswrapper[5025]: I1004 10:46:10.848237 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5bv89"] Oct 04 10:46:20 crc kubenswrapper[5025]: I1004 10:46:20.739410 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g2ggs" Oct 04 10:46:33 crc kubenswrapper[5025]: I1004 10:46:33.973916 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb"] Oct 04 10:46:33 crc kubenswrapper[5025]: I1004 10:46:33.976152 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:33 crc kubenswrapper[5025]: I1004 10:46:33.978114 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 04 10:46:33 crc kubenswrapper[5025]: I1004 10:46:33.990455 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb"] Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.065640 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.065832 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7m2w\" (UniqueName: \"kubernetes.io/projected/6257e369-9d92-403e-85cf-3990b895da8d-kube-api-access-f7m2w\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.066297 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.167422 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.167755 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.167801 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7m2w\" (UniqueName: \"kubernetes.io/projected/6257e369-9d92-403e-85cf-3990b895da8d-kube-api-access-f7m2w\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.168651 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.169127 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.197692 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7m2w\" (UniqueName: \"kubernetes.io/projected/6257e369-9d92-403e-85cf-3990b895da8d-kube-api-access-f7m2w\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.303525 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.571633 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb"] Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.946657 5025 generic.go:334] "Generic (PLEG): container finished" podID="6257e369-9d92-403e-85cf-3990b895da8d" containerID="49303e27b1277a136814579a60acfaee6c45895182cbf996d9dd6a05757cc656" exitCode=0 Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.946706 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" event={"ID":"6257e369-9d92-403e-85cf-3990b895da8d","Type":"ContainerDied","Data":"49303e27b1277a136814579a60acfaee6c45895182cbf996d9dd6a05757cc656"} Oct 04 10:46:34 crc kubenswrapper[5025]: I1004 10:46:34.946730 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" event={"ID":"6257e369-9d92-403e-85cf-3990b895da8d","Type":"ContainerStarted","Data":"419e527cf7ad530495b968105c900ce564125782c7f2b23c4f0156d6d88a74b3"} Oct 04 10:46:35 crc kubenswrapper[5025]: I1004 10:46:35.894218 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5bv89" podUID="0f85106f-c448-496a-8199-9a03990115ac" containerName="console" containerID="cri-o://4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b" gracePeriod=15 Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.259276 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5bv89_0f85106f-c448-496a-8199-9a03990115ac/console/0.log" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.259351 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.394844 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr97x\" (UniqueName: \"kubernetes.io/projected/0f85106f-c448-496a-8199-9a03990115ac-kube-api-access-pr97x\") pod \"0f85106f-c448-496a-8199-9a03990115ac\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.395296 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-oauth-config\") pod \"0f85106f-c448-496a-8199-9a03990115ac\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.395387 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-service-ca\") pod \"0f85106f-c448-496a-8199-9a03990115ac\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.395422 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-serving-cert\") pod \"0f85106f-c448-496a-8199-9a03990115ac\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.395451 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-trusted-ca-bundle\") pod \"0f85106f-c448-496a-8199-9a03990115ac\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.395491 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-console-config\") pod \"0f85106f-c448-496a-8199-9a03990115ac\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.395526 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-oauth-serving-cert\") pod \"0f85106f-c448-496a-8199-9a03990115ac\" (UID: \"0f85106f-c448-496a-8199-9a03990115ac\") " Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.395994 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-service-ca" (OuterVolumeSpecName: "service-ca") pod "0f85106f-c448-496a-8199-9a03990115ac" (UID: "0f85106f-c448-496a-8199-9a03990115ac"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.396595 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "0f85106f-c448-496a-8199-9a03990115ac" (UID: "0f85106f-c448-496a-8199-9a03990115ac"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.396644 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-console-config" (OuterVolumeSpecName: "console-config") pod "0f85106f-c448-496a-8199-9a03990115ac" (UID: "0f85106f-c448-496a-8199-9a03990115ac"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.396838 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "0f85106f-c448-496a-8199-9a03990115ac" (UID: "0f85106f-c448-496a-8199-9a03990115ac"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.401816 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "0f85106f-c448-496a-8199-9a03990115ac" (UID: "0f85106f-c448-496a-8199-9a03990115ac"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.402612 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "0f85106f-c448-496a-8199-9a03990115ac" (UID: "0f85106f-c448-496a-8199-9a03990115ac"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.404416 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f85106f-c448-496a-8199-9a03990115ac-kube-api-access-pr97x" (OuterVolumeSpecName: "kube-api-access-pr97x") pod "0f85106f-c448-496a-8199-9a03990115ac" (UID: "0f85106f-c448-496a-8199-9a03990115ac"). InnerVolumeSpecName "kube-api-access-pr97x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.497346 5025 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-service-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.497388 5025 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.497406 5025 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.497420 5025 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-console-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.497434 5025 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0f85106f-c448-496a-8199-9a03990115ac-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.497450 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr97x\" (UniqueName: \"kubernetes.io/projected/0f85106f-c448-496a-8199-9a03990115ac-kube-api-access-pr97x\") on node \"crc\" DevicePath \"\"" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.497466 5025 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0f85106f-c448-496a-8199-9a03990115ac-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.961193 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5bv89_0f85106f-c448-496a-8199-9a03990115ac/console/0.log" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.961249 5025 generic.go:334] "Generic (PLEG): container finished" podID="0f85106f-c448-496a-8199-9a03990115ac" containerID="4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b" exitCode=2 Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.961304 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5bv89" event={"ID":"0f85106f-c448-496a-8199-9a03990115ac","Type":"ContainerDied","Data":"4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b"} Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.961347 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5bv89" event={"ID":"0f85106f-c448-496a-8199-9a03990115ac","Type":"ContainerDied","Data":"0de25330acbe04083b83abfbb8f0d993ed949bbe0e4b4582627162b9c6e8053f"} Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.961367 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5bv89" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.961440 5025 scope.go:117] "RemoveContainer" containerID="4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.964672 5025 generic.go:334] "Generic (PLEG): container finished" podID="6257e369-9d92-403e-85cf-3990b895da8d" containerID="80c0d09d0241ac184615ef1c1dd5b6e369c3a87c89f26f24958150ca6ad1d910" exitCode=0 Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.964723 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" event={"ID":"6257e369-9d92-403e-85cf-3990b895da8d","Type":"ContainerDied","Data":"80c0d09d0241ac184615ef1c1dd5b6e369c3a87c89f26f24958150ca6ad1d910"} Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.987078 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5bv89"] Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.990822 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5bv89"] Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.991413 5025 scope.go:117] "RemoveContainer" containerID="4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b" Oct 04 10:46:36 crc kubenswrapper[5025]: E1004 10:46:36.991925 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b\": container with ID starting with 4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b not found: ID does not exist" containerID="4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b" Oct 04 10:46:36 crc kubenswrapper[5025]: I1004 10:46:36.991969 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b"} err="failed to get container status \"4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b\": rpc error: code = NotFound desc = could not find container \"4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b\": container with ID starting with 4a467e6cddef53e661fccff12161bd8121d12d58429a77abb683143baa4a346b not found: ID does not exist" Oct 04 10:46:37 crc kubenswrapper[5025]: I1004 10:46:37.975976 5025 generic.go:334] "Generic (PLEG): container finished" podID="6257e369-9d92-403e-85cf-3990b895da8d" containerID="c75b3c85aed96a668473f7eea991435de1853cf26db38e9e38492d88d69c23ed" exitCode=0 Oct 04 10:46:37 crc kubenswrapper[5025]: I1004 10:46:37.976093 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" event={"ID":"6257e369-9d92-403e-85cf-3990b895da8d","Type":"ContainerDied","Data":"c75b3c85aed96a668473f7eea991435de1853cf26db38e9e38492d88d69c23ed"} Oct 04 10:46:38 crc kubenswrapper[5025]: I1004 10:46:38.425373 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f85106f-c448-496a-8199-9a03990115ac" path="/var/lib/kubelet/pods/0f85106f-c448-496a-8199-9a03990115ac/volumes" Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.265323 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.437939 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-util\") pod \"6257e369-9d92-403e-85cf-3990b895da8d\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.438540 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-bundle\") pod \"6257e369-9d92-403e-85cf-3990b895da8d\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.438612 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7m2w\" (UniqueName: \"kubernetes.io/projected/6257e369-9d92-403e-85cf-3990b895da8d-kube-api-access-f7m2w\") pod \"6257e369-9d92-403e-85cf-3990b895da8d\" (UID: \"6257e369-9d92-403e-85cf-3990b895da8d\") " Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.440057 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-bundle" (OuterVolumeSpecName: "bundle") pod "6257e369-9d92-403e-85cf-3990b895da8d" (UID: "6257e369-9d92-403e-85cf-3990b895da8d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.445395 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6257e369-9d92-403e-85cf-3990b895da8d-kube-api-access-f7m2w" (OuterVolumeSpecName: "kube-api-access-f7m2w") pod "6257e369-9d92-403e-85cf-3990b895da8d" (UID: "6257e369-9d92-403e-85cf-3990b895da8d"). InnerVolumeSpecName "kube-api-access-f7m2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.457359 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-util" (OuterVolumeSpecName: "util") pod "6257e369-9d92-403e-85cf-3990b895da8d" (UID: "6257e369-9d92-403e-85cf-3990b895da8d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.541420 5025 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-util\") on node \"crc\" DevicePath \"\"" Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.541623 5025 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6257e369-9d92-403e-85cf-3990b895da8d-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.541686 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7m2w\" (UniqueName: \"kubernetes.io/projected/6257e369-9d92-403e-85cf-3990b895da8d-kube-api-access-f7m2w\") on node \"crc\" DevicePath \"\"" Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.993744 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" event={"ID":"6257e369-9d92-403e-85cf-3990b895da8d","Type":"ContainerDied","Data":"419e527cf7ad530495b968105c900ce564125782c7f2b23c4f0156d6d88a74b3"} Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.993799 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="419e527cf7ad530495b968105c900ce564125782c7f2b23c4f0156d6d88a74b3" Oct 04 10:46:39 crc kubenswrapper[5025]: I1004 10:46:39.993882 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.816522 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs"] Oct 04 10:46:48 crc kubenswrapper[5025]: E1004 10:46:48.817212 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f85106f-c448-496a-8199-9a03990115ac" containerName="console" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.817226 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f85106f-c448-496a-8199-9a03990115ac" containerName="console" Oct 04 10:46:48 crc kubenswrapper[5025]: E1004 10:46:48.817246 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6257e369-9d92-403e-85cf-3990b895da8d" containerName="extract" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.817253 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="6257e369-9d92-403e-85cf-3990b895da8d" containerName="extract" Oct 04 10:46:48 crc kubenswrapper[5025]: E1004 10:46:48.817265 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6257e369-9d92-403e-85cf-3990b895da8d" containerName="util" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.817272 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="6257e369-9d92-403e-85cf-3990b895da8d" containerName="util" Oct 04 10:46:48 crc kubenswrapper[5025]: E1004 10:46:48.817284 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6257e369-9d92-403e-85cf-3990b895da8d" containerName="pull" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.817291 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="6257e369-9d92-403e-85cf-3990b895da8d" containerName="pull" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.817424 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="6257e369-9d92-403e-85cf-3990b895da8d" containerName="extract" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.817445 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f85106f-c448-496a-8199-9a03990115ac" containerName="console" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.817872 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.822679 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.822771 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.823133 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-gwqm7" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.823416 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.828102 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.863167 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs"] Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.971155 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvqg4\" (UniqueName: \"kubernetes.io/projected/b87c8f6a-58fe-47e2-822b-ad21c79d0123-kube-api-access-wvqg4\") pod \"metallb-operator-controller-manager-66f54ffcff-9lmgs\" (UID: \"b87c8f6a-58fe-47e2-822b-ad21c79d0123\") " pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.971212 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b87c8f6a-58fe-47e2-822b-ad21c79d0123-webhook-cert\") pod \"metallb-operator-controller-manager-66f54ffcff-9lmgs\" (UID: \"b87c8f6a-58fe-47e2-822b-ad21c79d0123\") " pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:48 crc kubenswrapper[5025]: I1004 10:46:48.971264 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b87c8f6a-58fe-47e2-822b-ad21c79d0123-apiservice-cert\") pod \"metallb-operator-controller-manager-66f54ffcff-9lmgs\" (UID: \"b87c8f6a-58fe-47e2-822b-ad21c79d0123\") " pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.072644 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b87c8f6a-58fe-47e2-822b-ad21c79d0123-apiservice-cert\") pod \"metallb-operator-controller-manager-66f54ffcff-9lmgs\" (UID: \"b87c8f6a-58fe-47e2-822b-ad21c79d0123\") " pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.072749 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvqg4\" (UniqueName: \"kubernetes.io/projected/b87c8f6a-58fe-47e2-822b-ad21c79d0123-kube-api-access-wvqg4\") pod \"metallb-operator-controller-manager-66f54ffcff-9lmgs\" (UID: \"b87c8f6a-58fe-47e2-822b-ad21c79d0123\") " pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.072781 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b87c8f6a-58fe-47e2-822b-ad21c79d0123-webhook-cert\") pod \"metallb-operator-controller-manager-66f54ffcff-9lmgs\" (UID: \"b87c8f6a-58fe-47e2-822b-ad21c79d0123\") " pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.074907 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9"] Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.075770 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.078351 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.078579 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-5k22l" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.078843 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b87c8f6a-58fe-47e2-822b-ad21c79d0123-apiservice-cert\") pod \"metallb-operator-controller-manager-66f54ffcff-9lmgs\" (UID: \"b87c8f6a-58fe-47e2-822b-ad21c79d0123\") " pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.079702 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b87c8f6a-58fe-47e2-822b-ad21c79d0123-webhook-cert\") pod \"metallb-operator-controller-manager-66f54ffcff-9lmgs\" (UID: \"b87c8f6a-58fe-47e2-822b-ad21c79d0123\") " pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.092620 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvqg4\" (UniqueName: \"kubernetes.io/projected/b87c8f6a-58fe-47e2-822b-ad21c79d0123-kube-api-access-wvqg4\") pod \"metallb-operator-controller-manager-66f54ffcff-9lmgs\" (UID: \"b87c8f6a-58fe-47e2-822b-ad21c79d0123\") " pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.093870 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.100535 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9"] Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.136398 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.174051 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kts5\" (UniqueName: \"kubernetes.io/projected/20275cb8-c871-4219-8aee-4f47f917ed72-kube-api-access-4kts5\") pod \"metallb-operator-webhook-server-764d6c8955-dhmw9\" (UID: \"20275cb8-c871-4219-8aee-4f47f917ed72\") " pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.174114 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20275cb8-c871-4219-8aee-4f47f917ed72-webhook-cert\") pod \"metallb-operator-webhook-server-764d6c8955-dhmw9\" (UID: \"20275cb8-c871-4219-8aee-4f47f917ed72\") " pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.174178 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20275cb8-c871-4219-8aee-4f47f917ed72-apiservice-cert\") pod \"metallb-operator-webhook-server-764d6c8955-dhmw9\" (UID: \"20275cb8-c871-4219-8aee-4f47f917ed72\") " pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.274938 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kts5\" (UniqueName: \"kubernetes.io/projected/20275cb8-c871-4219-8aee-4f47f917ed72-kube-api-access-4kts5\") pod \"metallb-operator-webhook-server-764d6c8955-dhmw9\" (UID: \"20275cb8-c871-4219-8aee-4f47f917ed72\") " pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.275231 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20275cb8-c871-4219-8aee-4f47f917ed72-webhook-cert\") pod \"metallb-operator-webhook-server-764d6c8955-dhmw9\" (UID: \"20275cb8-c871-4219-8aee-4f47f917ed72\") " pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.275290 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20275cb8-c871-4219-8aee-4f47f917ed72-apiservice-cert\") pod \"metallb-operator-webhook-server-764d6c8955-dhmw9\" (UID: \"20275cb8-c871-4219-8aee-4f47f917ed72\") " pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.282498 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/20275cb8-c871-4219-8aee-4f47f917ed72-apiservice-cert\") pod \"metallb-operator-webhook-server-764d6c8955-dhmw9\" (UID: \"20275cb8-c871-4219-8aee-4f47f917ed72\") " pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.282973 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/20275cb8-c871-4219-8aee-4f47f917ed72-webhook-cert\") pod \"metallb-operator-webhook-server-764d6c8955-dhmw9\" (UID: \"20275cb8-c871-4219-8aee-4f47f917ed72\") " pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.305044 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kts5\" (UniqueName: \"kubernetes.io/projected/20275cb8-c871-4219-8aee-4f47f917ed72-kube-api-access-4kts5\") pod \"metallb-operator-webhook-server-764d6c8955-dhmw9\" (UID: \"20275cb8-c871-4219-8aee-4f47f917ed72\") " pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.388676 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs"] Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.426222 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:49 crc kubenswrapper[5025]: I1004 10:46:49.652301 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9"] Oct 04 10:46:49 crc kubenswrapper[5025]: W1004 10:46:49.660367 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20275cb8_c871_4219_8aee_4f47f917ed72.slice/crio-00a738e0ccf5dde489615e16cae0c4c53f092b4ecc5641df9fc490ca892e2df4 WatchSource:0}: Error finding container 00a738e0ccf5dde489615e16cae0c4c53f092b4ecc5641df9fc490ca892e2df4: Status 404 returned error can't find the container with id 00a738e0ccf5dde489615e16cae0c4c53f092b4ecc5641df9fc490ca892e2df4 Oct 04 10:46:50 crc kubenswrapper[5025]: I1004 10:46:50.042122 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" event={"ID":"b87c8f6a-58fe-47e2-822b-ad21c79d0123","Type":"ContainerStarted","Data":"d863a7db985da9ba41f282ffeca5325f898ebc6fff5166c4f08a109d06a5bb9a"} Oct 04 10:46:50 crc kubenswrapper[5025]: I1004 10:46:50.043255 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" event={"ID":"20275cb8-c871-4219-8aee-4f47f917ed72","Type":"ContainerStarted","Data":"00a738e0ccf5dde489615e16cae0c4c53f092b4ecc5641df9fc490ca892e2df4"} Oct 04 10:46:54 crc kubenswrapper[5025]: I1004 10:46:54.068705 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" event={"ID":"b87c8f6a-58fe-47e2-822b-ad21c79d0123","Type":"ContainerStarted","Data":"3d455cedccd65f7b6e61d2cf6244265ea8e2d67ff2d7013e43db21f33421df42"} Oct 04 10:46:54 crc kubenswrapper[5025]: I1004 10:46:54.069138 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:46:54 crc kubenswrapper[5025]: I1004 10:46:54.070707 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" event={"ID":"20275cb8-c871-4219-8aee-4f47f917ed72","Type":"ContainerStarted","Data":"4d3db87f8e10336b2ab173a42e618e00312d5806f1c17531f26609ef13b4df3c"} Oct 04 10:46:54 crc kubenswrapper[5025]: I1004 10:46:54.070851 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:46:54 crc kubenswrapper[5025]: I1004 10:46:54.085320 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" podStartSLOduration=3.180970756 podStartE2EDuration="6.085307482s" podCreationTimestamp="2025-10-04 10:46:48 +0000 UTC" firstStartedPulling="2025-10-04 10:46:49.39674657 +0000 UTC m=+737.821713450" lastFinishedPulling="2025-10-04 10:46:52.301083296 +0000 UTC m=+740.726050176" observedRunningTime="2025-10-04 10:46:54.083764137 +0000 UTC m=+742.508731017" watchObservedRunningTime="2025-10-04 10:46:54.085307482 +0000 UTC m=+742.510274362" Oct 04 10:46:54 crc kubenswrapper[5025]: I1004 10:46:54.116091 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" podStartSLOduration=1.038318862 podStartE2EDuration="5.116077092s" podCreationTimestamp="2025-10-04 10:46:49 +0000 UTC" firstStartedPulling="2025-10-04 10:46:49.663215275 +0000 UTC m=+738.088182155" lastFinishedPulling="2025-10-04 10:46:53.740973505 +0000 UTC m=+742.165940385" observedRunningTime="2025-10-04 10:46:54.114055663 +0000 UTC m=+742.539022543" watchObservedRunningTime="2025-10-04 10:46:54.116077092 +0000 UTC m=+742.541043972" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.240816 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-w42lh"] Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.242101 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" podUID="1877d566-a1e5-424c-87db-cb04d1fdc1b6" containerName="controller-manager" containerID="cri-o://7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493" gracePeriod=30 Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.317954 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn"] Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.318222 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" podUID="9ada2393-a845-408e-b527-e26c7af5b037" containerName="route-controller-manager" containerID="cri-o://91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1" gracePeriod=30 Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.669619 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.677368 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.827524 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-client-ca\") pod \"9ada2393-a845-408e-b527-e26c7af5b037\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.827629 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ada2393-a845-408e-b527-e26c7af5b037-serving-cert\") pod \"9ada2393-a845-408e-b527-e26c7af5b037\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.827706 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1877d566-a1e5-424c-87db-cb04d1fdc1b6-serving-cert\") pod \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.827731 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68hm6\" (UniqueName: \"kubernetes.io/projected/9ada2393-a845-408e-b527-e26c7af5b037-kube-api-access-68hm6\") pod \"9ada2393-a845-408e-b527-e26c7af5b037\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.827767 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9d6z\" (UniqueName: \"kubernetes.io/projected/1877d566-a1e5-424c-87db-cb04d1fdc1b6-kube-api-access-s9d6z\") pod \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.827803 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-config\") pod \"9ada2393-a845-408e-b527-e26c7af5b037\" (UID: \"9ada2393-a845-408e-b527-e26c7af5b037\") " Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.827828 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-proxy-ca-bundles\") pod \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.827863 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-client-ca\") pod \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.827891 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-config\") pod \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\" (UID: \"1877d566-a1e5-424c-87db-cb04d1fdc1b6\") " Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.828307 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-client-ca" (OuterVolumeSpecName: "client-ca") pod "9ada2393-a845-408e-b527-e26c7af5b037" (UID: "9ada2393-a845-408e-b527-e26c7af5b037"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.828370 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-client-ca" (OuterVolumeSpecName: "client-ca") pod "1877d566-a1e5-424c-87db-cb04d1fdc1b6" (UID: "1877d566-a1e5-424c-87db-cb04d1fdc1b6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.828488 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1877d566-a1e5-424c-87db-cb04d1fdc1b6" (UID: "1877d566-a1e5-424c-87db-cb04d1fdc1b6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.828481 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-config" (OuterVolumeSpecName: "config") pod "9ada2393-a845-408e-b527-e26c7af5b037" (UID: "9ada2393-a845-408e-b527-e26c7af5b037"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.828823 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-config" (OuterVolumeSpecName: "config") pod "1877d566-a1e5-424c-87db-cb04d1fdc1b6" (UID: "1877d566-a1e5-424c-87db-cb04d1fdc1b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.828848 5025 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.828870 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ada2393-a845-408e-b527-e26c7af5b037-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.828881 5025 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.828894 5025 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-client-ca\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.849192 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1877d566-a1e5-424c-87db-cb04d1fdc1b6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1877d566-a1e5-424c-87db-cb04d1fdc1b6" (UID: "1877d566-a1e5-424c-87db-cb04d1fdc1b6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.849305 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1877d566-a1e5-424c-87db-cb04d1fdc1b6-kube-api-access-s9d6z" (OuterVolumeSpecName: "kube-api-access-s9d6z") pod "1877d566-a1e5-424c-87db-cb04d1fdc1b6" (UID: "1877d566-a1e5-424c-87db-cb04d1fdc1b6"). InnerVolumeSpecName "kube-api-access-s9d6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.849771 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ada2393-a845-408e-b527-e26c7af5b037-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9ada2393-a845-408e-b527-e26c7af5b037" (UID: "9ada2393-a845-408e-b527-e26c7af5b037"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.864477 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ada2393-a845-408e-b527-e26c7af5b037-kube-api-access-68hm6" (OuterVolumeSpecName: "kube-api-access-68hm6") pod "9ada2393-a845-408e-b527-e26c7af5b037" (UID: "9ada2393-a845-408e-b527-e26c7af5b037"). InnerVolumeSpecName "kube-api-access-68hm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.929990 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1877d566-a1e5-424c-87db-cb04d1fdc1b6-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.930036 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68hm6\" (UniqueName: \"kubernetes.io/projected/9ada2393-a845-408e-b527-e26c7af5b037-kube-api-access-68hm6\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.930048 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9d6z\" (UniqueName: \"kubernetes.io/projected/1877d566-a1e5-424c-87db-cb04d1fdc1b6-kube-api-access-s9d6z\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.930056 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1877d566-a1e5-424c-87db-cb04d1fdc1b6-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:00 crc kubenswrapper[5025]: I1004 10:47:00.930065 5025 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9ada2393-a845-408e-b527-e26c7af5b037-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.119136 5025 generic.go:334] "Generic (PLEG): container finished" podID="9ada2393-a845-408e-b527-e26c7af5b037" containerID="91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1" exitCode=0 Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.119200 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.119212 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" event={"ID":"9ada2393-a845-408e-b527-e26c7af5b037","Type":"ContainerDied","Data":"91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1"} Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.119269 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn" event={"ID":"9ada2393-a845-408e-b527-e26c7af5b037","Type":"ContainerDied","Data":"29de14d693ac83733e88730a5fee52883e5522fcb32389e929f99684c31f333b"} Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.119314 5025 scope.go:117] "RemoveContainer" containerID="91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.120857 5025 generic.go:334] "Generic (PLEG): container finished" podID="1877d566-a1e5-424c-87db-cb04d1fdc1b6" containerID="7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493" exitCode=0 Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.120887 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" event={"ID":"1877d566-a1e5-424c-87db-cb04d1fdc1b6","Type":"ContainerDied","Data":"7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493"} Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.120903 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.120911 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-w42lh" event={"ID":"1877d566-a1e5-424c-87db-cb04d1fdc1b6","Type":"ContainerDied","Data":"f37da1a433cfb303b3a65ddb705b304e04d2f5ba5fc6e5397cca454401467cab"} Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.140372 5025 scope.go:117] "RemoveContainer" containerID="91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1" Oct 04 10:47:01 crc kubenswrapper[5025]: E1004 10:47:01.140805 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1\": container with ID starting with 91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1 not found: ID does not exist" containerID="91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.140873 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1"} err="failed to get container status \"91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1\": rpc error: code = NotFound desc = could not find container \"91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1\": container with ID starting with 91e7ec5419b5cbc82756e4bc4f30d84206cc17d1eb3a7fe0f431bed1cd85b8d1 not found: ID does not exist" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.140921 5025 scope.go:117] "RemoveContainer" containerID="7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.150106 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn"] Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.158881 5025 scope.go:117] "RemoveContainer" containerID="7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493" Oct 04 10:47:01 crc kubenswrapper[5025]: E1004 10:47:01.159362 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493\": container with ID starting with 7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493 not found: ID does not exist" containerID="7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.159406 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493"} err="failed to get container status \"7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493\": rpc error: code = NotFound desc = could not find container \"7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493\": container with ID starting with 7140bba26bbf207784694c116ed17d7c56f6c4378f664b1dd12819869f753493 not found: ID does not exist" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.160188 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rbctn"] Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.168165 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-w42lh"] Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.174621 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-w42lh"] Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.941800 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-789d55f67f-2wfj5"] Oct 04 10:47:01 crc kubenswrapper[5025]: E1004 10:47:01.942045 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1877d566-a1e5-424c-87db-cb04d1fdc1b6" containerName="controller-manager" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.942056 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="1877d566-a1e5-424c-87db-cb04d1fdc1b6" containerName="controller-manager" Oct 04 10:47:01 crc kubenswrapper[5025]: E1004 10:47:01.942066 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ada2393-a845-408e-b527-e26c7af5b037" containerName="route-controller-manager" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.942072 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ada2393-a845-408e-b527-e26c7af5b037" containerName="route-controller-manager" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.942159 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="1877d566-a1e5-424c-87db-cb04d1fdc1b6" containerName="controller-manager" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.942172 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ada2393-a845-408e-b527-e26c7af5b037" containerName="route-controller-manager" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.942536 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.944646 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.945156 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.945245 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.945334 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.945558 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.945594 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.954351 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.960596 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-789d55f67f-2wfj5"] Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.980104 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9"] Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.980970 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.986205 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.986497 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.986665 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.986694 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.986777 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 04 10:47:01 crc kubenswrapper[5025]: I1004 10:47:01.987520 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.004348 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9"] Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.042419 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ghcw\" (UniqueName: \"kubernetes.io/projected/6979506a-4559-4475-8d20-785c74855e5d-kube-api-access-8ghcw\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.042729 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6979506a-4559-4475-8d20-785c74855e5d-serving-cert\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.042847 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6979506a-4559-4475-8d20-785c74855e5d-config\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.042944 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6979506a-4559-4475-8d20-785c74855e5d-proxy-ca-bundles\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.043113 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6979506a-4559-4475-8d20-785c74855e5d-client-ca\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.144524 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/189a2c64-8309-4c7e-9b52-22c579b606d8-serving-cert\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.144936 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6979506a-4559-4475-8d20-785c74855e5d-client-ca\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.146148 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ghcw\" (UniqueName: \"kubernetes.io/projected/6979506a-4559-4475-8d20-785c74855e5d-kube-api-access-8ghcw\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.146289 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6979506a-4559-4475-8d20-785c74855e5d-serving-cert\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.146411 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6979506a-4559-4475-8d20-785c74855e5d-config\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.146519 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6979506a-4559-4475-8d20-785c74855e5d-proxy-ca-bundles\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.146644 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/189a2c64-8309-4c7e-9b52-22c579b606d8-client-ca\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.146739 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189a2c64-8309-4c7e-9b52-22c579b606d8-config\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.146842 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67wrq\" (UniqueName: \"kubernetes.io/projected/189a2c64-8309-4c7e-9b52-22c579b606d8-kube-api-access-67wrq\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.148082 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6979506a-4559-4475-8d20-785c74855e5d-proxy-ca-bundles\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.146080 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6979506a-4559-4475-8d20-785c74855e5d-client-ca\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.148713 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6979506a-4559-4475-8d20-785c74855e5d-config\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.167056 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6979506a-4559-4475-8d20-785c74855e5d-serving-cert\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.183725 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ghcw\" (UniqueName: \"kubernetes.io/projected/6979506a-4559-4475-8d20-785c74855e5d-kube-api-access-8ghcw\") pod \"controller-manager-789d55f67f-2wfj5\" (UID: \"6979506a-4559-4475-8d20-785c74855e5d\") " pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.247997 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/189a2c64-8309-4c7e-9b52-22c579b606d8-client-ca\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.248057 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189a2c64-8309-4c7e-9b52-22c579b606d8-config\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.248079 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67wrq\" (UniqueName: \"kubernetes.io/projected/189a2c64-8309-4c7e-9b52-22c579b606d8-kube-api-access-67wrq\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.248103 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/189a2c64-8309-4c7e-9b52-22c579b606d8-serving-cert\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.248963 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/189a2c64-8309-4c7e-9b52-22c579b606d8-client-ca\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.249473 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/189a2c64-8309-4c7e-9b52-22c579b606d8-config\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.251601 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/189a2c64-8309-4c7e-9b52-22c579b606d8-serving-cert\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.265647 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.271826 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67wrq\" (UniqueName: \"kubernetes.io/projected/189a2c64-8309-4c7e-9b52-22c579b606d8-kube-api-access-67wrq\") pod \"route-controller-manager-6c5585cc9c-86mr9\" (UID: \"189a2c64-8309-4c7e-9b52-22c579b606d8\") " pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.304352 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.421181 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1877d566-a1e5-424c-87db-cb04d1fdc1b6" path="/var/lib/kubelet/pods/1877d566-a1e5-424c-87db-cb04d1fdc1b6/volumes" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.421907 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ada2393-a845-408e-b527-e26c7af5b037" path="/var/lib/kubelet/pods/9ada2393-a845-408e-b527-e26c7af5b037/volumes" Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.519292 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-789d55f67f-2wfj5"] Oct 04 10:47:02 crc kubenswrapper[5025]: I1004 10:47:02.764292 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9"] Oct 04 10:47:02 crc kubenswrapper[5025]: W1004 10:47:02.774212 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod189a2c64_8309_4c7e_9b52_22c579b606d8.slice/crio-25bc01664a65e71aab318e75347a09ee2880208f6da405f0ed4c4f4a0e75066b WatchSource:0}: Error finding container 25bc01664a65e71aab318e75347a09ee2880208f6da405f0ed4c4f4a0e75066b: Status 404 returned error can't find the container with id 25bc01664a65e71aab318e75347a09ee2880208f6da405f0ed4c4f4a0e75066b Oct 04 10:47:03 crc kubenswrapper[5025]: I1004 10:47:03.133723 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" event={"ID":"189a2c64-8309-4c7e-9b52-22c579b606d8","Type":"ContainerStarted","Data":"ffafce78b28715ad30487656add8da7b142b3e87f46ee26c9a0a761eb2f57273"} Oct 04 10:47:03 crc kubenswrapper[5025]: I1004 10:47:03.135247 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:03 crc kubenswrapper[5025]: I1004 10:47:03.135375 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" event={"ID":"189a2c64-8309-4c7e-9b52-22c579b606d8","Type":"ContainerStarted","Data":"25bc01664a65e71aab318e75347a09ee2880208f6da405f0ed4c4f4a0e75066b"} Oct 04 10:47:03 crc kubenswrapper[5025]: I1004 10:47:03.135468 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" event={"ID":"6979506a-4559-4475-8d20-785c74855e5d","Type":"ContainerStarted","Data":"875f3b5c4b67030403885b9af035bc6c5907a0f91526dd7f9d4da692ab02ca64"} Oct 04 10:47:03 crc kubenswrapper[5025]: I1004 10:47:03.135541 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" event={"ID":"6979506a-4559-4475-8d20-785c74855e5d","Type":"ContainerStarted","Data":"b4a70ebfd0378cbf6efc7be1f2f1528127fdc4038c9ec78a0dc44bbd75fe142d"} Oct 04 10:47:03 crc kubenswrapper[5025]: I1004 10:47:03.135610 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:03 crc kubenswrapper[5025]: I1004 10:47:03.149954 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" Oct 04 10:47:03 crc kubenswrapper[5025]: I1004 10:47:03.154826 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" podStartSLOduration=2.15481071 podStartE2EDuration="2.15481071s" podCreationTimestamp="2025-10-04 10:47:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:47:03.152319588 +0000 UTC m=+751.577286468" watchObservedRunningTime="2025-10-04 10:47:03.15481071 +0000 UTC m=+751.579777590" Oct 04 10:47:03 crc kubenswrapper[5025]: I1004 10:47:03.172292 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-789d55f67f-2wfj5" podStartSLOduration=2.172268225 podStartE2EDuration="2.172268225s" podCreationTimestamp="2025-10-04 10:47:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:47:03.169130774 +0000 UTC m=+751.594097664" watchObservedRunningTime="2025-10-04 10:47:03.172268225 +0000 UTC m=+751.597235105" Oct 04 10:47:03 crc kubenswrapper[5025]: I1004 10:47:03.441175 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6c5585cc9c-86mr9" Oct 04 10:47:06 crc kubenswrapper[5025]: I1004 10:47:06.642753 5025 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 04 10:47:09 crc kubenswrapper[5025]: I1004 10:47:09.449271 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-764d6c8955-dhmw9" Oct 04 10:47:14 crc kubenswrapper[5025]: I1004 10:47:14.713391 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:47:14 crc kubenswrapper[5025]: I1004 10:47:14.714164 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:47:29 crc kubenswrapper[5025]: I1004 10:47:29.138601 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-66f54ffcff-9lmgs" Oct 04 10:47:29 crc kubenswrapper[5025]: I1004 10:47:29.978801 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6"] Oct 04 10:47:29 crc kubenswrapper[5025]: I1004 10:47:29.979694 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" Oct 04 10:47:29 crc kubenswrapper[5025]: I1004 10:47:29.982448 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-7rrhp"] Oct 04 10:47:29 crc kubenswrapper[5025]: I1004 10:47:29.986547 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 04 10:47:29 crc kubenswrapper[5025]: I1004 10:47:29.987756 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:29 crc kubenswrapper[5025]: I1004 10:47:29.987977 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-vd9mx" Oct 04 10:47:29 crc kubenswrapper[5025]: I1004 10:47:29.991353 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 04 10:47:29 crc kubenswrapper[5025]: I1004 10:47:29.992552 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 04 10:47:29 crc kubenswrapper[5025]: I1004 10:47:29.992692 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6"] Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.066817 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6rf4m"] Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.067922 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.070532 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-65nk4" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.070568 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.070541 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.070927 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.094939 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-jqv9k"] Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.095905 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.097778 5025 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.109117 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-jqv9k"] Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.118058 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-frr-startup\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.118108 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-metrics-certs\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.118135 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-frr-sockets\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.118237 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79-cert\") pod \"frr-k8s-webhook-server-64bf5d555-ptxm6\" (UID: \"27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.118315 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-reloader\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.118388 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zhkl\" (UniqueName: \"kubernetes.io/projected/27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79-kube-api-access-4zhkl\") pod \"frr-k8s-webhook-server-64bf5d555-ptxm6\" (UID: \"27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.118452 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-frr-conf\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.118499 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-metrics\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.118537 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrpnx\" (UniqueName: \"kubernetes.io/projected/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-kube-api-access-rrpnx\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.219966 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-reloader\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220047 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkxvz\" (UniqueName: \"kubernetes.io/projected/d9c990ef-7b47-4ece-b38d-822182cd7407-kube-api-access-lkxvz\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220093 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zhkl\" (UniqueName: \"kubernetes.io/projected/27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79-kube-api-access-4zhkl\") pod \"frr-k8s-webhook-server-64bf5d555-ptxm6\" (UID: \"27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220139 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-frr-conf\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220201 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-metrics\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220231 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrpnx\" (UniqueName: \"kubernetes.io/projected/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-kube-api-access-rrpnx\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220268 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24bfff08-7460-44c2-8358-c18faa7933fd-metrics-certs\") pod \"controller-68d546b9d8-jqv9k\" (UID: \"24bfff08-7460-44c2-8358-c18faa7933fd\") " pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220292 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-memberlist\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220325 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-frr-startup\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220395 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-metrics-certs\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220433 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq7dc\" (UniqueName: \"kubernetes.io/projected/24bfff08-7460-44c2-8358-c18faa7933fd-kube-api-access-lq7dc\") pod \"controller-68d546b9d8-jqv9k\" (UID: \"24bfff08-7460-44c2-8358-c18faa7933fd\") " pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220463 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-frr-sockets\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220493 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24bfff08-7460-44c2-8358-c18faa7933fd-cert\") pod \"controller-68d546b9d8-jqv9k\" (UID: \"24bfff08-7460-44c2-8358-c18faa7933fd\") " pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: E1004 10:47:30.220502 5025 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Oct 04 10:47:30 crc kubenswrapper[5025]: E1004 10:47:30.220585 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-metrics-certs podName:3aa58e9d-ce3d-4f94-aec2-aa1d870d309b nodeName:}" failed. No retries permitted until 2025-10-04 10:47:30.720561375 +0000 UTC m=+779.145528315 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-metrics-certs") pod "frr-k8s-7rrhp" (UID: "3aa58e9d-ce3d-4f94-aec2-aa1d870d309b") : secret "frr-k8s-certs-secret" not found Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220516 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d9c990ef-7b47-4ece-b38d-822182cd7407-metallb-excludel2\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220685 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79-cert\") pod \"frr-k8s-webhook-server-64bf5d555-ptxm6\" (UID: \"27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.220738 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-metrics-certs\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.221092 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-reloader\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.221092 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-frr-conf\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.221158 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-metrics\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.221239 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-frr-sockets\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.221796 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-frr-startup\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.227727 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79-cert\") pod \"frr-k8s-webhook-server-64bf5d555-ptxm6\" (UID: \"27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.235182 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrpnx\" (UniqueName: \"kubernetes.io/projected/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-kube-api-access-rrpnx\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.237873 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zhkl\" (UniqueName: \"kubernetes.io/projected/27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79-kube-api-access-4zhkl\") pod \"frr-k8s-webhook-server-64bf5d555-ptxm6\" (UID: \"27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.310415 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.322091 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq7dc\" (UniqueName: \"kubernetes.io/projected/24bfff08-7460-44c2-8358-c18faa7933fd-kube-api-access-lq7dc\") pod \"controller-68d546b9d8-jqv9k\" (UID: \"24bfff08-7460-44c2-8358-c18faa7933fd\") " pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.322511 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24bfff08-7460-44c2-8358-c18faa7933fd-cert\") pod \"controller-68d546b9d8-jqv9k\" (UID: \"24bfff08-7460-44c2-8358-c18faa7933fd\") " pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.322649 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d9c990ef-7b47-4ece-b38d-822182cd7407-metallb-excludel2\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.322774 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-metrics-certs\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.322888 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkxvz\" (UniqueName: \"kubernetes.io/projected/d9c990ef-7b47-4ece-b38d-822182cd7407-kube-api-access-lkxvz\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.323128 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24bfff08-7460-44c2-8358-c18faa7933fd-metrics-certs\") pod \"controller-68d546b9d8-jqv9k\" (UID: \"24bfff08-7460-44c2-8358-c18faa7933fd\") " pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.323276 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-memberlist\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.323363 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d9c990ef-7b47-4ece-b38d-822182cd7407-metallb-excludel2\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: E1004 10:47:30.323705 5025 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 10:47:30 crc kubenswrapper[5025]: E1004 10:47:30.323772 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-memberlist podName:d9c990ef-7b47-4ece-b38d-822182cd7407 nodeName:}" failed. No retries permitted until 2025-10-04 10:47:30.823754629 +0000 UTC m=+779.248721509 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-memberlist") pod "speaker-6rf4m" (UID: "d9c990ef-7b47-4ece-b38d-822182cd7407") : secret "metallb-memberlist" not found Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.326176 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-metrics-certs\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.330487 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/24bfff08-7460-44c2-8358-c18faa7933fd-cert\") pod \"controller-68d546b9d8-jqv9k\" (UID: \"24bfff08-7460-44c2-8358-c18faa7933fd\") " pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.330685 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/24bfff08-7460-44c2-8358-c18faa7933fd-metrics-certs\") pod \"controller-68d546b9d8-jqv9k\" (UID: \"24bfff08-7460-44c2-8358-c18faa7933fd\") " pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.339889 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkxvz\" (UniqueName: \"kubernetes.io/projected/d9c990ef-7b47-4ece-b38d-822182cd7407-kube-api-access-lkxvz\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.340216 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq7dc\" (UniqueName: \"kubernetes.io/projected/24bfff08-7460-44c2-8358-c18faa7933fd-kube-api-access-lq7dc\") pod \"controller-68d546b9d8-jqv9k\" (UID: \"24bfff08-7460-44c2-8358-c18faa7933fd\") " pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.408846 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.729817 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-metrics-certs\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.738201 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3aa58e9d-ce3d-4f94-aec2-aa1d870d309b-metrics-certs\") pod \"frr-k8s-7rrhp\" (UID: \"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b\") " pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.743099 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6"] Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.831423 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-memberlist\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:30 crc kubenswrapper[5025]: E1004 10:47:30.831603 5025 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 04 10:47:30 crc kubenswrapper[5025]: E1004 10:47:30.831664 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-memberlist podName:d9c990ef-7b47-4ece-b38d-822182cd7407 nodeName:}" failed. No retries permitted until 2025-10-04 10:47:31.831649417 +0000 UTC m=+780.256616297 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-memberlist") pod "speaker-6rf4m" (UID: "d9c990ef-7b47-4ece-b38d-822182cd7407") : secret "metallb-memberlist" not found Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.835410 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-jqv9k"] Oct 04 10:47:30 crc kubenswrapper[5025]: W1004 10:47:30.843989 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24bfff08_7460_44c2_8358_c18faa7933fd.slice/crio-b134c50351e88a5ab034b8a993c1518b52f72079f1b8dd50183658db885a8780 WatchSource:0}: Error finding container b134c50351e88a5ab034b8a993c1518b52f72079f1b8dd50183658db885a8780: Status 404 returned error can't find the container with id b134c50351e88a5ab034b8a993c1518b52f72079f1b8dd50183658db885a8780 Oct 04 10:47:30 crc kubenswrapper[5025]: I1004 10:47:30.920701 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:31 crc kubenswrapper[5025]: I1004 10:47:31.308093 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" event={"ID":"27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79","Type":"ContainerStarted","Data":"65f5d31dd88d2487ce76e264e0d21cb8751797d0bfe0893b9a2c91115332e1c9"} Oct 04 10:47:31 crc kubenswrapper[5025]: I1004 10:47:31.309246 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rrhp" event={"ID":"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b","Type":"ContainerStarted","Data":"155e1f9a8517c7cc4c52b3563f8dabe222b5d6f9abbaaa706448a63494bdf223"} Oct 04 10:47:31 crc kubenswrapper[5025]: I1004 10:47:31.311281 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-jqv9k" event={"ID":"24bfff08-7460-44c2-8358-c18faa7933fd","Type":"ContainerStarted","Data":"f8dc333c76526a9de0262c4b89eac041aad571f4e93cd488f53834a0c357fc01"} Oct 04 10:47:31 crc kubenswrapper[5025]: I1004 10:47:31.311320 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-jqv9k" event={"ID":"24bfff08-7460-44c2-8358-c18faa7933fd","Type":"ContainerStarted","Data":"12c22c3fc53ef7e3ecc3ca22ce4e0c8889a7079a5e680509f4c80d89ea70706a"} Oct 04 10:47:31 crc kubenswrapper[5025]: I1004 10:47:31.311332 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-jqv9k" event={"ID":"24bfff08-7460-44c2-8358-c18faa7933fd","Type":"ContainerStarted","Data":"b134c50351e88a5ab034b8a993c1518b52f72079f1b8dd50183658db885a8780"} Oct 04 10:47:31 crc kubenswrapper[5025]: I1004 10:47:31.311431 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:31 crc kubenswrapper[5025]: I1004 10:47:31.329959 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-jqv9k" podStartSLOduration=1.329935115 podStartE2EDuration="1.329935115s" podCreationTimestamp="2025-10-04 10:47:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:47:31.3259459 +0000 UTC m=+779.750912780" watchObservedRunningTime="2025-10-04 10:47:31.329935115 +0000 UTC m=+779.754901995" Oct 04 10:47:31 crc kubenswrapper[5025]: I1004 10:47:31.848128 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-memberlist\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:31 crc kubenswrapper[5025]: I1004 10:47:31.874700 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d9c990ef-7b47-4ece-b38d-822182cd7407-memberlist\") pod \"speaker-6rf4m\" (UID: \"d9c990ef-7b47-4ece-b38d-822182cd7407\") " pod="metallb-system/speaker-6rf4m" Oct 04 10:47:31 crc kubenswrapper[5025]: I1004 10:47:31.883557 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6rf4m" Oct 04 10:47:32 crc kubenswrapper[5025]: I1004 10:47:32.329327 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6rf4m" event={"ID":"d9c990ef-7b47-4ece-b38d-822182cd7407","Type":"ContainerStarted","Data":"ccea82ee1dcbbd36c30db23b070019b29c70e2bb21399301deee61dcebb038b4"} Oct 04 10:47:32 crc kubenswrapper[5025]: I1004 10:47:32.329366 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6rf4m" event={"ID":"d9c990ef-7b47-4ece-b38d-822182cd7407","Type":"ContainerStarted","Data":"8c96179b9f43e49929974c6e4e8dfd4deddb6df56903e2edd781cce46fd3a8d4"} Oct 04 10:47:33 crc kubenswrapper[5025]: I1004 10:47:33.341302 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6rf4m" event={"ID":"d9c990ef-7b47-4ece-b38d-822182cd7407","Type":"ContainerStarted","Data":"f7890655efbea6668a79f5c5b81bf38563213f26b10d24bab6a0e01a5b5991c7"} Oct 04 10:47:33 crc kubenswrapper[5025]: I1004 10:47:33.341475 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6rf4m" Oct 04 10:47:33 crc kubenswrapper[5025]: I1004 10:47:33.370486 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6rf4m" podStartSLOduration=3.370465123 podStartE2EDuration="3.370465123s" podCreationTimestamp="2025-10-04 10:47:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:47:33.366484707 +0000 UTC m=+781.791451607" watchObservedRunningTime="2025-10-04 10:47:33.370465123 +0000 UTC m=+781.795432003" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.260084 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mtzkf"] Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.262638 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.273466 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtzkf"] Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.432888 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-utilities\") pod \"redhat-marketplace-mtzkf\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.432944 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-catalog-content\") pod \"redhat-marketplace-mtzkf\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.433028 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swttx\" (UniqueName: \"kubernetes.io/projected/df70c71e-26ac-4d62-9de4-1dd67ba3642a-kube-api-access-swttx\") pod \"redhat-marketplace-mtzkf\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.534077 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swttx\" (UniqueName: \"kubernetes.io/projected/df70c71e-26ac-4d62-9de4-1dd67ba3642a-kube-api-access-swttx\") pod \"redhat-marketplace-mtzkf\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.534202 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-utilities\") pod \"redhat-marketplace-mtzkf\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.534224 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-catalog-content\") pod \"redhat-marketplace-mtzkf\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.534621 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-catalog-content\") pod \"redhat-marketplace-mtzkf\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.534919 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-utilities\") pod \"redhat-marketplace-mtzkf\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.577753 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swttx\" (UniqueName: \"kubernetes.io/projected/df70c71e-26ac-4d62-9de4-1dd67ba3642a-kube-api-access-swttx\") pod \"redhat-marketplace-mtzkf\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.584464 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:37 crc kubenswrapper[5025]: I1004 10:47:37.998655 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtzkf"] Oct 04 10:47:38 crc kubenswrapper[5025]: I1004 10:47:38.377987 5025 generic.go:334] "Generic (PLEG): container finished" podID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerID="76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415" exitCode=0 Oct 04 10:47:38 crc kubenswrapper[5025]: I1004 10:47:38.378067 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtzkf" event={"ID":"df70c71e-26ac-4d62-9de4-1dd67ba3642a","Type":"ContainerDied","Data":"76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415"} Oct 04 10:47:38 crc kubenswrapper[5025]: I1004 10:47:38.378126 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtzkf" event={"ID":"df70c71e-26ac-4d62-9de4-1dd67ba3642a","Type":"ContainerStarted","Data":"bfa0d5f900a0639cbcc7502d253c0949f4d3e68f863fcbd5210529e687cd217a"} Oct 04 10:47:38 crc kubenswrapper[5025]: I1004 10:47:38.381143 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" event={"ID":"27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79","Type":"ContainerStarted","Data":"fe01a41fbdb917ebbc143e08b784237788115c4a5b704a60ad680e8062842748"} Oct 04 10:47:38 crc kubenswrapper[5025]: I1004 10:47:38.381340 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" Oct 04 10:47:38 crc kubenswrapper[5025]: I1004 10:47:38.386695 5025 generic.go:334] "Generic (PLEG): container finished" podID="3aa58e9d-ce3d-4f94-aec2-aa1d870d309b" containerID="0ee9d061a7c66fd28ea0bb44992310580d386e996dac2472b80b0659980fe77f" exitCode=0 Oct 04 10:47:38 crc kubenswrapper[5025]: I1004 10:47:38.386794 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rrhp" event={"ID":"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b","Type":"ContainerDied","Data":"0ee9d061a7c66fd28ea0bb44992310580d386e996dac2472b80b0659980fe77f"} Oct 04 10:47:39 crc kubenswrapper[5025]: I1004 10:47:39.395059 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rrhp" event={"ID":"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b","Type":"ContainerDied","Data":"14b1e4ded97940a109e900c528a1bf313021a4cb8323f67b7421f0b720b8b5ec"} Oct 04 10:47:39 crc kubenswrapper[5025]: I1004 10:47:39.396092 5025 generic.go:334] "Generic (PLEG): container finished" podID="3aa58e9d-ce3d-4f94-aec2-aa1d870d309b" containerID="14b1e4ded97940a109e900c528a1bf313021a4cb8323f67b7421f0b720b8b5ec" exitCode=0 Oct 04 10:47:39 crc kubenswrapper[5025]: I1004 10:47:39.433053 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" podStartSLOduration=3.610451575 podStartE2EDuration="10.433034627s" podCreationTimestamp="2025-10-04 10:47:29 +0000 UTC" firstStartedPulling="2025-10-04 10:47:30.751914101 +0000 UTC m=+779.176880981" lastFinishedPulling="2025-10-04 10:47:37.574497153 +0000 UTC m=+785.999464033" observedRunningTime="2025-10-04 10:47:38.456443326 +0000 UTC m=+786.881410206" watchObservedRunningTime="2025-10-04 10:47:39.433034627 +0000 UTC m=+787.858001507" Oct 04 10:47:40 crc kubenswrapper[5025]: I1004 10:47:40.407114 5025 generic.go:334] "Generic (PLEG): container finished" podID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerID="65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c" exitCode=0 Oct 04 10:47:40 crc kubenswrapper[5025]: I1004 10:47:40.407190 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtzkf" event={"ID":"df70c71e-26ac-4d62-9de4-1dd67ba3642a","Type":"ContainerDied","Data":"65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c"} Oct 04 10:47:40 crc kubenswrapper[5025]: I1004 10:47:40.413250 5025 generic.go:334] "Generic (PLEG): container finished" podID="3aa58e9d-ce3d-4f94-aec2-aa1d870d309b" containerID="9e1032fd4daa638c3fbabe8ffcc15e2663bacd11acf0105bfad50cefbb369c23" exitCode=0 Oct 04 10:47:40 crc kubenswrapper[5025]: I1004 10:47:40.424757 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-jqv9k" Oct 04 10:47:40 crc kubenswrapper[5025]: I1004 10:47:40.424820 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rrhp" event={"ID":"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b","Type":"ContainerDied","Data":"9e1032fd4daa638c3fbabe8ffcc15e2663bacd11acf0105bfad50cefbb369c23"} Oct 04 10:47:41 crc kubenswrapper[5025]: I1004 10:47:41.428578 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtzkf" event={"ID":"df70c71e-26ac-4d62-9de4-1dd67ba3642a","Type":"ContainerStarted","Data":"56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b"} Oct 04 10:47:41 crc kubenswrapper[5025]: I1004 10:47:41.436207 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rrhp" event={"ID":"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b","Type":"ContainerStarted","Data":"7fc94a0614e6dc4c6b884d461029aeac595badd751c53698f6fabddab8430f2c"} Oct 04 10:47:41 crc kubenswrapper[5025]: I1004 10:47:41.436289 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rrhp" event={"ID":"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b","Type":"ContainerStarted","Data":"1e190396772772e339e28a5869f4f0f488ecabdc5df260903bfb1585c2e5d677"} Oct 04 10:47:41 crc kubenswrapper[5025]: I1004 10:47:41.436308 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rrhp" event={"ID":"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b","Type":"ContainerStarted","Data":"4f4ce8fc3ecc5f162403cc10cf4ce9eb44f0b9de5ab95daee44a0105e81fe552"} Oct 04 10:47:41 crc kubenswrapper[5025]: I1004 10:47:41.436324 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rrhp" event={"ID":"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b","Type":"ContainerStarted","Data":"b0304c63c2d96a8113b152d4baa116e7819110d52308c715c97506be0da7b127"} Oct 04 10:47:41 crc kubenswrapper[5025]: I1004 10:47:41.448628 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mtzkf" podStartSLOduration=1.9376732620000001 podStartE2EDuration="4.448606543s" podCreationTimestamp="2025-10-04 10:47:37 +0000 UTC" firstStartedPulling="2025-10-04 10:47:38.380260933 +0000 UTC m=+786.805227853" lastFinishedPulling="2025-10-04 10:47:40.891194234 +0000 UTC m=+789.316161134" observedRunningTime="2025-10-04 10:47:41.447096 +0000 UTC m=+789.872062900" watchObservedRunningTime="2025-10-04 10:47:41.448606543 +0000 UTC m=+789.873573433" Oct 04 10:47:42 crc kubenswrapper[5025]: I1004 10:47:42.449451 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rrhp" event={"ID":"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b","Type":"ContainerStarted","Data":"73f7000374ce4b2d164981a4501f5e71b2af3b4fe414d3ac39767ada3f529615"} Oct 04 10:47:42 crc kubenswrapper[5025]: I1004 10:47:42.449509 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-7rrhp" event={"ID":"3aa58e9d-ce3d-4f94-aec2-aa1d870d309b","Type":"ContainerStarted","Data":"dfc62a4400f643b13dae10ff016a28542b7574e4661128d9b50e320d191f7ce7"} Oct 04 10:47:42 crc kubenswrapper[5025]: I1004 10:47:42.450047 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:42 crc kubenswrapper[5025]: I1004 10:47:42.485714 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-7rrhp" podStartSLOduration=6.921357968 podStartE2EDuration="13.485687212s" podCreationTimestamp="2025-10-04 10:47:29 +0000 UTC" firstStartedPulling="2025-10-04 10:47:31.036725477 +0000 UTC m=+779.461692357" lastFinishedPulling="2025-10-04 10:47:37.601054721 +0000 UTC m=+786.026021601" observedRunningTime="2025-10-04 10:47:42.483292573 +0000 UTC m=+790.908259493" watchObservedRunningTime="2025-10-04 10:47:42.485687212 +0000 UTC m=+790.910654132" Oct 04 10:47:44 crc kubenswrapper[5025]: I1004 10:47:44.714079 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:47:44 crc kubenswrapper[5025]: I1004 10:47:44.714514 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:47:45 crc kubenswrapper[5025]: I1004 10:47:45.921795 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:45 crc kubenswrapper[5025]: I1004 10:47:45.987937 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.206202 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q87c7"] Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.217608 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.224729 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q87c7"] Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.393139 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf9zx\" (UniqueName: \"kubernetes.io/projected/82a371e2-9754-4a14-b3cf-e8fb10e1d480-kube-api-access-gf9zx\") pod \"redhat-operators-q87c7\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.393243 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-utilities\") pod \"redhat-operators-q87c7\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.393276 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-catalog-content\") pod \"redhat-operators-q87c7\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.494740 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf9zx\" (UniqueName: \"kubernetes.io/projected/82a371e2-9754-4a14-b3cf-e8fb10e1d480-kube-api-access-gf9zx\") pod \"redhat-operators-q87c7\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.494884 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-utilities\") pod \"redhat-operators-q87c7\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.494907 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-catalog-content\") pod \"redhat-operators-q87c7\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.495348 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-utilities\") pod \"redhat-operators-q87c7\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.495426 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-catalog-content\") pod \"redhat-operators-q87c7\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.520978 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf9zx\" (UniqueName: \"kubernetes.io/projected/82a371e2-9754-4a14-b3cf-e8fb10e1d480-kube-api-access-gf9zx\") pod \"redhat-operators-q87c7\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.554643 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.585540 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.585591 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.622358 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:47 crc kubenswrapper[5025]: I1004 10:47:47.972055 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q87c7"] Oct 04 10:47:47 crc kubenswrapper[5025]: W1004 10:47:47.977172 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82a371e2_9754_4a14_b3cf_e8fb10e1d480.slice/crio-f1b60beca225ddf6591950bb0e39303ee4884d8f9295173de09167258d92cf0d WatchSource:0}: Error finding container f1b60beca225ddf6591950bb0e39303ee4884d8f9295173de09167258d92cf0d: Status 404 returned error can't find the container with id f1b60beca225ddf6591950bb0e39303ee4884d8f9295173de09167258d92cf0d Oct 04 10:47:48 crc kubenswrapper[5025]: I1004 10:47:48.497689 5025 generic.go:334] "Generic (PLEG): container finished" podID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerID="d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4" exitCode=0 Oct 04 10:47:48 crc kubenswrapper[5025]: I1004 10:47:48.497761 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q87c7" event={"ID":"82a371e2-9754-4a14-b3cf-e8fb10e1d480","Type":"ContainerDied","Data":"d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4"} Oct 04 10:47:48 crc kubenswrapper[5025]: I1004 10:47:48.498131 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q87c7" event={"ID":"82a371e2-9754-4a14-b3cf-e8fb10e1d480","Type":"ContainerStarted","Data":"f1b60beca225ddf6591950bb0e39303ee4884d8f9295173de09167258d92cf0d"} Oct 04 10:47:48 crc kubenswrapper[5025]: I1004 10:47:48.573267 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:49 crc kubenswrapper[5025]: I1004 10:47:49.957326 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtzkf"] Oct 04 10:47:50 crc kubenswrapper[5025]: I1004 10:47:50.316783 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-ptxm6" Oct 04 10:47:50 crc kubenswrapper[5025]: I1004 10:47:50.513551 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mtzkf" podUID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerName="registry-server" containerID="cri-o://56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b" gracePeriod=2 Oct 04 10:47:50 crc kubenswrapper[5025]: I1004 10:47:50.923500 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-7rrhp" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.266330 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.450841 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swttx\" (UniqueName: \"kubernetes.io/projected/df70c71e-26ac-4d62-9de4-1dd67ba3642a-kube-api-access-swttx\") pod \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.451076 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-catalog-content\") pod \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.451195 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-utilities\") pod \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\" (UID: \"df70c71e-26ac-4d62-9de4-1dd67ba3642a\") " Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.452629 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-utilities" (OuterVolumeSpecName: "utilities") pod "df70c71e-26ac-4d62-9de4-1dd67ba3642a" (UID: "df70c71e-26ac-4d62-9de4-1dd67ba3642a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.459543 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df70c71e-26ac-4d62-9de4-1dd67ba3642a-kube-api-access-swttx" (OuterVolumeSpecName: "kube-api-access-swttx") pod "df70c71e-26ac-4d62-9de4-1dd67ba3642a" (UID: "df70c71e-26ac-4d62-9de4-1dd67ba3642a"). InnerVolumeSpecName "kube-api-access-swttx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.472829 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df70c71e-26ac-4d62-9de4-1dd67ba3642a" (UID: "df70c71e-26ac-4d62-9de4-1dd67ba3642a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.522263 5025 generic.go:334] "Generic (PLEG): container finished" podID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerID="56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b" exitCode=0 Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.522305 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtzkf" event={"ID":"df70c71e-26ac-4d62-9de4-1dd67ba3642a","Type":"ContainerDied","Data":"56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b"} Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.522329 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mtzkf" event={"ID":"df70c71e-26ac-4d62-9de4-1dd67ba3642a","Type":"ContainerDied","Data":"bfa0d5f900a0639cbcc7502d253c0949f4d3e68f863fcbd5210529e687cd217a"} Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.522362 5025 scope.go:117] "RemoveContainer" containerID="56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.522401 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mtzkf" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.552824 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swttx\" (UniqueName: \"kubernetes.io/projected/df70c71e-26ac-4d62-9de4-1dd67ba3642a-kube-api-access-swttx\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.552869 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.552884 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df70c71e-26ac-4d62-9de4-1dd67ba3642a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.558675 5025 scope.go:117] "RemoveContainer" containerID="65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.564222 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtzkf"] Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.566879 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mtzkf"] Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.590705 5025 scope.go:117] "RemoveContainer" containerID="76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.612326 5025 scope.go:117] "RemoveContainer" containerID="56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b" Oct 04 10:47:51 crc kubenswrapper[5025]: E1004 10:47:51.616348 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b\": container with ID starting with 56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b not found: ID does not exist" containerID="56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.616424 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b"} err="failed to get container status \"56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b\": rpc error: code = NotFound desc = could not find container \"56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b\": container with ID starting with 56221a25cf662e42edb050c06cf4d8eb81308477fa4165b11b8a10a3dac52e6b not found: ID does not exist" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.616469 5025 scope.go:117] "RemoveContainer" containerID="65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c" Oct 04 10:47:51 crc kubenswrapper[5025]: E1004 10:47:51.616946 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c\": container with ID starting with 65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c not found: ID does not exist" containerID="65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.616978 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c"} err="failed to get container status \"65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c\": rpc error: code = NotFound desc = could not find container \"65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c\": container with ID starting with 65d2bb9bf19b3396d101fd1eb663f28d197341c200bb141239bd7892fb46a93c not found: ID does not exist" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.616998 5025 scope.go:117] "RemoveContainer" containerID="76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415" Oct 04 10:47:51 crc kubenswrapper[5025]: E1004 10:47:51.617374 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415\": container with ID starting with 76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415 not found: ID does not exist" containerID="76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.617424 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415"} err="failed to get container status \"76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415\": rpc error: code = NotFound desc = could not find container \"76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415\": container with ID starting with 76fa2cc8f8d4f0e90e9fe46a14b95ef255e3b69f66718a2dd54b35bb731bb415 not found: ID does not exist" Oct 04 10:47:51 crc kubenswrapper[5025]: I1004 10:47:51.887944 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6rf4m" Oct 04 10:47:52 crc kubenswrapper[5025]: I1004 10:47:52.422530 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" path="/var/lib/kubelet/pods/df70c71e-26ac-4d62-9de4-1dd67ba3642a/volumes" Oct 04 10:47:52 crc kubenswrapper[5025]: I1004 10:47:52.534628 5025 generic.go:334] "Generic (PLEG): container finished" podID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerID="48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7" exitCode=0 Oct 04 10:47:52 crc kubenswrapper[5025]: I1004 10:47:52.534707 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q87c7" event={"ID":"82a371e2-9754-4a14-b3cf-e8fb10e1d480","Type":"ContainerDied","Data":"48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7"} Oct 04 10:47:53 crc kubenswrapper[5025]: I1004 10:47:53.544384 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q87c7" event={"ID":"82a371e2-9754-4a14-b3cf-e8fb10e1d480","Type":"ContainerStarted","Data":"79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8"} Oct 04 10:47:53 crc kubenswrapper[5025]: I1004 10:47:53.566641 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q87c7" podStartSLOduration=2.085168684 podStartE2EDuration="6.566619556s" podCreationTimestamp="2025-10-04 10:47:47 +0000 UTC" firstStartedPulling="2025-10-04 10:47:48.500661761 +0000 UTC m=+796.925628651" lastFinishedPulling="2025-10-04 10:47:52.982112623 +0000 UTC m=+801.407079523" observedRunningTime="2025-10-04 10:47:53.561272331 +0000 UTC m=+801.986239241" watchObservedRunningTime="2025-10-04 10:47:53.566619556 +0000 UTC m=+801.991586476" Oct 04 10:47:57 crc kubenswrapper[5025]: I1004 10:47:57.555663 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:57 crc kubenswrapper[5025]: I1004 10:47:57.556140 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.163727 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-jjmsm"] Oct 04 10:47:58 crc kubenswrapper[5025]: E1004 10:47:58.164063 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerName="extract-utilities" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.164084 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerName="extract-utilities" Oct 04 10:47:58 crc kubenswrapper[5025]: E1004 10:47:58.164098 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerName="extract-content" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.164106 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerName="extract-content" Oct 04 10:47:58 crc kubenswrapper[5025]: E1004 10:47:58.164122 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerName="registry-server" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.164131 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerName="registry-server" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.164271 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="df70c71e-26ac-4d62-9de4-1dd67ba3642a" containerName="registry-server" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.164797 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jjmsm" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.169753 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-czvq4" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.169789 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.170123 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.175315 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jjmsm"] Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.345773 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbjxk\" (UniqueName: \"kubernetes.io/projected/f3962134-5912-472b-903d-29f945e216e6-kube-api-access-rbjxk\") pod \"openstack-operator-index-jjmsm\" (UID: \"f3962134-5912-472b-903d-29f945e216e6\") " pod="openstack-operators/openstack-operator-index-jjmsm" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.447485 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbjxk\" (UniqueName: \"kubernetes.io/projected/f3962134-5912-472b-903d-29f945e216e6-kube-api-access-rbjxk\") pod \"openstack-operator-index-jjmsm\" (UID: \"f3962134-5912-472b-903d-29f945e216e6\") " pod="openstack-operators/openstack-operator-index-jjmsm" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.481106 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbjxk\" (UniqueName: \"kubernetes.io/projected/f3962134-5912-472b-903d-29f945e216e6-kube-api-access-rbjxk\") pod \"openstack-operator-index-jjmsm\" (UID: \"f3962134-5912-472b-903d-29f945e216e6\") " pod="openstack-operators/openstack-operator-index-jjmsm" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.486826 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-jjmsm" Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.596635 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-q87c7" podUID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerName="registry-server" probeResult="failure" output=< Oct 04 10:47:58 crc kubenswrapper[5025]: timeout: failed to connect service ":50051" within 1s Oct 04 10:47:58 crc kubenswrapper[5025]: > Oct 04 10:47:58 crc kubenswrapper[5025]: I1004 10:47:58.933760 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-jjmsm"] Oct 04 10:47:59 crc kubenswrapper[5025]: I1004 10:47:59.598777 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jjmsm" event={"ID":"f3962134-5912-472b-903d-29f945e216e6","Type":"ContainerStarted","Data":"3b488058c54829f23161550c67d858181039cc34ceb1a1abdb7195827efcfe15"} Oct 04 10:48:04 crc kubenswrapper[5025]: I1004 10:48:04.651230 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-jjmsm" event={"ID":"f3962134-5912-472b-903d-29f945e216e6","Type":"ContainerStarted","Data":"9ab6dbf21d3f6afbcfe09865ff1f088eec7e825ddbc571a0b4e60d051ec4c69c"} Oct 04 10:48:04 crc kubenswrapper[5025]: I1004 10:48:04.671406 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-jjmsm" podStartSLOduration=1.4301060429999999 podStartE2EDuration="6.671382858s" podCreationTimestamp="2025-10-04 10:47:58 +0000 UTC" firstStartedPulling="2025-10-04 10:47:58.946424106 +0000 UTC m=+807.371390976" lastFinishedPulling="2025-10-04 10:48:04.187700911 +0000 UTC m=+812.612667791" observedRunningTime="2025-10-04 10:48:04.669519944 +0000 UTC m=+813.094486844" watchObservedRunningTime="2025-10-04 10:48:04.671382858 +0000 UTC m=+813.096349768" Oct 04 10:48:07 crc kubenswrapper[5025]: I1004 10:48:07.623884 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:48:07 crc kubenswrapper[5025]: I1004 10:48:07.667501 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:48:08 crc kubenswrapper[5025]: I1004 10:48:08.487244 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-jjmsm" Oct 04 10:48:08 crc kubenswrapper[5025]: I1004 10:48:08.487281 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-jjmsm" Oct 04 10:48:08 crc kubenswrapper[5025]: I1004 10:48:08.532500 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-jjmsm" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.151822 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q87c7"] Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.152617 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q87c7" podUID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerName="registry-server" containerID="cri-o://79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8" gracePeriod=2 Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.556084 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.681790 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf9zx\" (UniqueName: \"kubernetes.io/projected/82a371e2-9754-4a14-b3cf-e8fb10e1d480-kube-api-access-gf9zx\") pod \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.682198 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-utilities\") pod \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.682239 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-catalog-content\") pod \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\" (UID: \"82a371e2-9754-4a14-b3cf-e8fb10e1d480\") " Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.682904 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-utilities" (OuterVolumeSpecName: "utilities") pod "82a371e2-9754-4a14-b3cf-e8fb10e1d480" (UID: "82a371e2-9754-4a14-b3cf-e8fb10e1d480"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.687160 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82a371e2-9754-4a14-b3cf-e8fb10e1d480-kube-api-access-gf9zx" (OuterVolumeSpecName: "kube-api-access-gf9zx") pod "82a371e2-9754-4a14-b3cf-e8fb10e1d480" (UID: "82a371e2-9754-4a14-b3cf-e8fb10e1d480"). InnerVolumeSpecName "kube-api-access-gf9zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.698402 5025 generic.go:334] "Generic (PLEG): container finished" podID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerID="79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8" exitCode=0 Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.698448 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q87c7" event={"ID":"82a371e2-9754-4a14-b3cf-e8fb10e1d480","Type":"ContainerDied","Data":"79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8"} Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.698473 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q87c7" event={"ID":"82a371e2-9754-4a14-b3cf-e8fb10e1d480","Type":"ContainerDied","Data":"f1b60beca225ddf6591950bb0e39303ee4884d8f9295173de09167258d92cf0d"} Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.698489 5025 scope.go:117] "RemoveContainer" containerID="79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.698612 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q87c7" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.721092 5025 scope.go:117] "RemoveContainer" containerID="48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.741673 5025 scope.go:117] "RemoveContainer" containerID="d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.761650 5025 scope.go:117] "RemoveContainer" containerID="79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8" Oct 04 10:48:11 crc kubenswrapper[5025]: E1004 10:48:11.762052 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8\": container with ID starting with 79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8 not found: ID does not exist" containerID="79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.762113 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8"} err="failed to get container status \"79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8\": rpc error: code = NotFound desc = could not find container \"79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8\": container with ID starting with 79828ee8dfb0ccf1f63149714de424205d6313113de2c2e4da005af5ba49dfe8 not found: ID does not exist" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.762144 5025 scope.go:117] "RemoveContainer" containerID="48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7" Oct 04 10:48:11 crc kubenswrapper[5025]: E1004 10:48:11.762418 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7\": container with ID starting with 48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7 not found: ID does not exist" containerID="48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.762450 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7"} err="failed to get container status \"48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7\": rpc error: code = NotFound desc = could not find container \"48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7\": container with ID starting with 48806e2ef189ab011f1f71846e3de6df9bc55f7e525975c2cf0efa0e3c5cffe7 not found: ID does not exist" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.762474 5025 scope.go:117] "RemoveContainer" containerID="d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4" Oct 04 10:48:11 crc kubenswrapper[5025]: E1004 10:48:11.762760 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4\": container with ID starting with d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4 not found: ID does not exist" containerID="d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.762811 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4"} err="failed to get container status \"d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4\": rpc error: code = NotFound desc = could not find container \"d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4\": container with ID starting with d83b15067b87890ca260ece14092818e9993e7340d70f139b4c3ba3953468da4 not found: ID does not exist" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.771709 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82a371e2-9754-4a14-b3cf-e8fb10e1d480" (UID: "82a371e2-9754-4a14-b3cf-e8fb10e1d480"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.784134 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf9zx\" (UniqueName: \"kubernetes.io/projected/82a371e2-9754-4a14-b3cf-e8fb10e1d480-kube-api-access-gf9zx\") on node \"crc\" DevicePath \"\"" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.784184 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:48:11 crc kubenswrapper[5025]: I1004 10:48:11.784194 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82a371e2-9754-4a14-b3cf-e8fb10e1d480-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:48:12 crc kubenswrapper[5025]: I1004 10:48:12.028398 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q87c7"] Oct 04 10:48:12 crc kubenswrapper[5025]: I1004 10:48:12.032525 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q87c7"] Oct 04 10:48:12 crc kubenswrapper[5025]: I1004 10:48:12.419951 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" path="/var/lib/kubelet/pods/82a371e2-9754-4a14-b3cf-e8fb10e1d480/volumes" Oct 04 10:48:14 crc kubenswrapper[5025]: I1004 10:48:14.714153 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:48:14 crc kubenswrapper[5025]: I1004 10:48:14.714551 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:48:14 crc kubenswrapper[5025]: I1004 10:48:14.714605 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:48:14 crc kubenswrapper[5025]: I1004 10:48:14.715350 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ca2d393da7fa0a921a67c473465c531835d20e5d5778a21a26fa2e7ee5204ae"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:48:14 crc kubenswrapper[5025]: I1004 10:48:14.715434 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://4ca2d393da7fa0a921a67c473465c531835d20e5d5778a21a26fa2e7ee5204ae" gracePeriod=600 Oct 04 10:48:15 crc kubenswrapper[5025]: I1004 10:48:15.735564 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="4ca2d393da7fa0a921a67c473465c531835d20e5d5778a21a26fa2e7ee5204ae" exitCode=0 Oct 04 10:48:15 crc kubenswrapper[5025]: I1004 10:48:15.735662 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"4ca2d393da7fa0a921a67c473465c531835d20e5d5778a21a26fa2e7ee5204ae"} Oct 04 10:48:15 crc kubenswrapper[5025]: I1004 10:48:15.736858 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"1fa2c044718960ddbde296231834fb7f5a1519308e0e7742cc723c16081ff583"} Oct 04 10:48:15 crc kubenswrapper[5025]: I1004 10:48:15.736905 5025 scope.go:117] "RemoveContainer" containerID="f00d677ff24dae96198611c79163ef5caff57e489eae259b2f02ea4b187f2545" Oct 04 10:48:18 crc kubenswrapper[5025]: I1004 10:48:18.523127 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-jjmsm" Oct 04 10:48:25 crc kubenswrapper[5025]: I1004 10:48:25.989144 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5"] Oct 04 10:48:25 crc kubenswrapper[5025]: E1004 10:48:25.990267 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerName="extract-utilities" Oct 04 10:48:25 crc kubenswrapper[5025]: I1004 10:48:25.990288 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerName="extract-utilities" Oct 04 10:48:25 crc kubenswrapper[5025]: E1004 10:48:25.990306 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerName="registry-server" Oct 04 10:48:25 crc kubenswrapper[5025]: I1004 10:48:25.990318 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerName="registry-server" Oct 04 10:48:25 crc kubenswrapper[5025]: E1004 10:48:25.990349 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerName="extract-content" Oct 04 10:48:25 crc kubenswrapper[5025]: I1004 10:48:25.990362 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerName="extract-content" Oct 04 10:48:25 crc kubenswrapper[5025]: I1004 10:48:25.990566 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="82a371e2-9754-4a14-b3cf-e8fb10e1d480" containerName="registry-server" Oct 04 10:48:25 crc kubenswrapper[5025]: I1004 10:48:25.991926 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:25 crc kubenswrapper[5025]: I1004 10:48:25.994565 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wvnf5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.005077 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5"] Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.176400 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz4fc\" (UniqueName: \"kubernetes.io/projected/c5ba0046-11c4-4e7a-ae23-ff2759a29332-kube-api-access-zz4fc\") pod \"69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.176451 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-util\") pod \"69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.176492 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-bundle\") pod \"69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.277928 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz4fc\" (UniqueName: \"kubernetes.io/projected/c5ba0046-11c4-4e7a-ae23-ff2759a29332-kube-api-access-zz4fc\") pod \"69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.278094 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-util\") pod \"69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.278212 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-bundle\") pod \"69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.279066 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-util\") pod \"69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.279233 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-bundle\") pod \"69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.298113 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz4fc\" (UniqueName: \"kubernetes.io/projected/c5ba0046-11c4-4e7a-ae23-ff2759a29332-kube-api-access-zz4fc\") pod \"69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.318901 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:26 crc kubenswrapper[5025]: I1004 10:48:26.806225 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5"] Oct 04 10:48:26 crc kubenswrapper[5025]: W1004 10:48:26.820500 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5ba0046_11c4_4e7a_ae23_ff2759a29332.slice/crio-b6b1b377d63bb18236ece8e37f096757e181fcfa2818b4dc07f9d851c0f05654 WatchSource:0}: Error finding container b6b1b377d63bb18236ece8e37f096757e181fcfa2818b4dc07f9d851c0f05654: Status 404 returned error can't find the container with id b6b1b377d63bb18236ece8e37f096757e181fcfa2818b4dc07f9d851c0f05654 Oct 04 10:48:27 crc kubenswrapper[5025]: I1004 10:48:27.831097 5025 generic.go:334] "Generic (PLEG): container finished" podID="c5ba0046-11c4-4e7a-ae23-ff2759a29332" containerID="a24bf63b6ce52a6c4fb5dafbebf8c93ccdef829a542f8aff218f67726ab9c00e" exitCode=0 Oct 04 10:48:27 crc kubenswrapper[5025]: I1004 10:48:27.834827 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" event={"ID":"c5ba0046-11c4-4e7a-ae23-ff2759a29332","Type":"ContainerDied","Data":"a24bf63b6ce52a6c4fb5dafbebf8c93ccdef829a542f8aff218f67726ab9c00e"} Oct 04 10:48:27 crc kubenswrapper[5025]: I1004 10:48:27.834922 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" event={"ID":"c5ba0046-11c4-4e7a-ae23-ff2759a29332","Type":"ContainerStarted","Data":"b6b1b377d63bb18236ece8e37f096757e181fcfa2818b4dc07f9d851c0f05654"} Oct 04 10:48:28 crc kubenswrapper[5025]: I1004 10:48:28.839878 5025 generic.go:334] "Generic (PLEG): container finished" podID="c5ba0046-11c4-4e7a-ae23-ff2759a29332" containerID="2aaa55f3f8568cb6f30879e69f8403182d247327d9c1af8b1d967975696ec748" exitCode=0 Oct 04 10:48:28 crc kubenswrapper[5025]: I1004 10:48:28.839980 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" event={"ID":"c5ba0046-11c4-4e7a-ae23-ff2759a29332","Type":"ContainerDied","Data":"2aaa55f3f8568cb6f30879e69f8403182d247327d9c1af8b1d967975696ec748"} Oct 04 10:48:29 crc kubenswrapper[5025]: I1004 10:48:29.851529 5025 generic.go:334] "Generic (PLEG): container finished" podID="c5ba0046-11c4-4e7a-ae23-ff2759a29332" containerID="860c22aa6cd745a05de0ef661b43e90495916436405fa3a66a23b8d10dd56924" exitCode=0 Oct 04 10:48:29 crc kubenswrapper[5025]: I1004 10:48:29.851586 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" event={"ID":"c5ba0046-11c4-4e7a-ae23-ff2759a29332","Type":"ContainerDied","Data":"860c22aa6cd745a05de0ef661b43e90495916436405fa3a66a23b8d10dd56924"} Oct 04 10:48:30 crc kubenswrapper[5025]: I1004 10:48:30.960294 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pz8s8"] Oct 04 10:48:30 crc kubenswrapper[5025]: I1004 10:48:30.962161 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:30 crc kubenswrapper[5025]: I1004 10:48:30.988486 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pz8s8"] Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.144044 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-utilities\") pod \"community-operators-pz8s8\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.144128 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-catalog-content\") pod \"community-operators-pz8s8\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.144205 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2lbr\" (UniqueName: \"kubernetes.io/projected/07edb013-6500-498e-9f66-f3d1b80f3c7b-kube-api-access-h2lbr\") pod \"community-operators-pz8s8\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.178511 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.245338 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-utilities\") pod \"community-operators-pz8s8\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.245409 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-catalog-content\") pod \"community-operators-pz8s8\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.245466 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2lbr\" (UniqueName: \"kubernetes.io/projected/07edb013-6500-498e-9f66-f3d1b80f3c7b-kube-api-access-h2lbr\") pod \"community-operators-pz8s8\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.245861 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-utilities\") pod \"community-operators-pz8s8\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.245915 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-catalog-content\") pod \"community-operators-pz8s8\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.276405 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2lbr\" (UniqueName: \"kubernetes.io/projected/07edb013-6500-498e-9f66-f3d1b80f3c7b-kube-api-access-h2lbr\") pod \"community-operators-pz8s8\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.284945 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.346718 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz4fc\" (UniqueName: \"kubernetes.io/projected/c5ba0046-11c4-4e7a-ae23-ff2759a29332-kube-api-access-zz4fc\") pod \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.347331 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-bundle\") pod \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.347456 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-util\") pod \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\" (UID: \"c5ba0046-11c4-4e7a-ae23-ff2759a29332\") " Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.348000 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-bundle" (OuterVolumeSpecName: "bundle") pod "c5ba0046-11c4-4e7a-ae23-ff2759a29332" (UID: "c5ba0046-11c4-4e7a-ae23-ff2759a29332"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.349242 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5ba0046-11c4-4e7a-ae23-ff2759a29332-kube-api-access-zz4fc" (OuterVolumeSpecName: "kube-api-access-zz4fc") pod "c5ba0046-11c4-4e7a-ae23-ff2759a29332" (UID: "c5ba0046-11c4-4e7a-ae23-ff2759a29332"). InnerVolumeSpecName "kube-api-access-zz4fc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.361970 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-util" (OuterVolumeSpecName: "util") pod "c5ba0046-11c4-4e7a-ae23-ff2759a29332" (UID: "c5ba0046-11c4-4e7a-ae23-ff2759a29332"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.449485 5025 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.449518 5025 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c5ba0046-11c4-4e7a-ae23-ff2759a29332-util\") on node \"crc\" DevicePath \"\"" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.449557 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz4fc\" (UniqueName: \"kubernetes.io/projected/c5ba0046-11c4-4e7a-ae23-ff2759a29332-kube-api-access-zz4fc\") on node \"crc\" DevicePath \"\"" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.533000 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pz8s8"] Oct 04 10:48:31 crc kubenswrapper[5025]: W1004 10:48:31.536073 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07edb013_6500_498e_9f66_f3d1b80f3c7b.slice/crio-d4a05bada3ccfd3b1276dbeba5a17b45bc19ab488a71ae2aeed8af7fb356f599 WatchSource:0}: Error finding container d4a05bada3ccfd3b1276dbeba5a17b45bc19ab488a71ae2aeed8af7fb356f599: Status 404 returned error can't find the container with id d4a05bada3ccfd3b1276dbeba5a17b45bc19ab488a71ae2aeed8af7fb356f599 Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.882213 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" event={"ID":"c5ba0046-11c4-4e7a-ae23-ff2759a29332","Type":"ContainerDied","Data":"b6b1b377d63bb18236ece8e37f096757e181fcfa2818b4dc07f9d851c0f05654"} Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.882629 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6b1b377d63bb18236ece8e37f096757e181fcfa2818b4dc07f9d851c0f05654" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.882281 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5" Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.884834 5025 generic.go:334] "Generic (PLEG): container finished" podID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerID="a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0" exitCode=0 Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.884877 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pz8s8" event={"ID":"07edb013-6500-498e-9f66-f3d1b80f3c7b","Type":"ContainerDied","Data":"a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0"} Oct 04 10:48:31 crc kubenswrapper[5025]: I1004 10:48:31.884904 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pz8s8" event={"ID":"07edb013-6500-498e-9f66-f3d1b80f3c7b","Type":"ContainerStarted","Data":"d4a05bada3ccfd3b1276dbeba5a17b45bc19ab488a71ae2aeed8af7fb356f599"} Oct 04 10:48:32 crc kubenswrapper[5025]: I1004 10:48:32.891251 5025 generic.go:334] "Generic (PLEG): container finished" podID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerID="176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88" exitCode=0 Oct 04 10:48:32 crc kubenswrapper[5025]: I1004 10:48:32.891330 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pz8s8" event={"ID":"07edb013-6500-498e-9f66-f3d1b80f3c7b","Type":"ContainerDied","Data":"176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88"} Oct 04 10:48:33 crc kubenswrapper[5025]: I1004 10:48:33.901645 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pz8s8" event={"ID":"07edb013-6500-498e-9f66-f3d1b80f3c7b","Type":"ContainerStarted","Data":"caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f"} Oct 04 10:48:33 crc kubenswrapper[5025]: I1004 10:48:33.928867 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pz8s8" podStartSLOduration=2.471639215 podStartE2EDuration="3.928846981s" podCreationTimestamp="2025-10-04 10:48:30 +0000 UTC" firstStartedPulling="2025-10-04 10:48:31.887357006 +0000 UTC m=+840.312323886" lastFinishedPulling="2025-10-04 10:48:33.344564772 +0000 UTC m=+841.769531652" observedRunningTime="2025-10-04 10:48:33.923162416 +0000 UTC m=+842.348129326" watchObservedRunningTime="2025-10-04 10:48:33.928846981 +0000 UTC m=+842.353813871" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.240998 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h"] Oct 04 10:48:38 crc kubenswrapper[5025]: E1004 10:48:38.241719 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ba0046-11c4-4e7a-ae23-ff2759a29332" containerName="pull" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.241730 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ba0046-11c4-4e7a-ae23-ff2759a29332" containerName="pull" Oct 04 10:48:38 crc kubenswrapper[5025]: E1004 10:48:38.241742 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ba0046-11c4-4e7a-ae23-ff2759a29332" containerName="extract" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.241748 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ba0046-11c4-4e7a-ae23-ff2759a29332" containerName="extract" Oct 04 10:48:38 crc kubenswrapper[5025]: E1004 10:48:38.241755 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ba0046-11c4-4e7a-ae23-ff2759a29332" containerName="util" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.241762 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ba0046-11c4-4e7a-ae23-ff2759a29332" containerName="util" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.241869 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ba0046-11c4-4e7a-ae23-ff2759a29332" containerName="extract" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.242437 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.243879 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-k8fhw" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.264065 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h"] Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.346314 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blqrw\" (UniqueName: \"kubernetes.io/projected/a1865149-283a-4933-b0d0-164afd0d1717-kube-api-access-blqrw\") pod \"openstack-operator-controller-operator-89c6d9c58-mtm9h\" (UID: \"a1865149-283a-4933-b0d0-164afd0d1717\") " pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.447056 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blqrw\" (UniqueName: \"kubernetes.io/projected/a1865149-283a-4933-b0d0-164afd0d1717-kube-api-access-blqrw\") pod \"openstack-operator-controller-operator-89c6d9c58-mtm9h\" (UID: \"a1865149-283a-4933-b0d0-164afd0d1717\") " pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.466817 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blqrw\" (UniqueName: \"kubernetes.io/projected/a1865149-283a-4933-b0d0-164afd0d1717-kube-api-access-blqrw\") pod \"openstack-operator-controller-operator-89c6d9c58-mtm9h\" (UID: \"a1865149-283a-4933-b0d0-164afd0d1717\") " pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.558420 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.840803 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h"] Oct 04 10:48:38 crc kubenswrapper[5025]: I1004 10:48:38.930440 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" event={"ID":"a1865149-283a-4933-b0d0-164afd0d1717","Type":"ContainerStarted","Data":"cd073184e4ff9260392d84085c386a2771a40d2550ad5d9f9f7da781bf91f43e"} Oct 04 10:48:41 crc kubenswrapper[5025]: I1004 10:48:41.285899 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:41 crc kubenswrapper[5025]: I1004 10:48:41.286237 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:41 crc kubenswrapper[5025]: I1004 10:48:41.336508 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:41 crc kubenswrapper[5025]: I1004 10:48:41.999488 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:42 crc kubenswrapper[5025]: I1004 10:48:42.976989 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" event={"ID":"a1865149-283a-4933-b0d0-164afd0d1717","Type":"ContainerStarted","Data":"7d5d5ab25d9b8b4f2c9b29687ffab78c51bb145dcdd301af0ac86a95e3ac229f"} Oct 04 10:48:43 crc kubenswrapper[5025]: I1004 10:48:43.740328 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pz8s8"] Oct 04 10:48:43 crc kubenswrapper[5025]: I1004 10:48:43.990104 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pz8s8" podUID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerName="registry-server" containerID="cri-o://caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f" gracePeriod=2 Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.703992 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.767178 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-utilities\") pod \"07edb013-6500-498e-9f66-f3d1b80f3c7b\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.767303 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2lbr\" (UniqueName: \"kubernetes.io/projected/07edb013-6500-498e-9f66-f3d1b80f3c7b-kube-api-access-h2lbr\") pod \"07edb013-6500-498e-9f66-f3d1b80f3c7b\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.767401 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-catalog-content\") pod \"07edb013-6500-498e-9f66-f3d1b80f3c7b\" (UID: \"07edb013-6500-498e-9f66-f3d1b80f3c7b\") " Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.768564 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-utilities" (OuterVolumeSpecName: "utilities") pod "07edb013-6500-498e-9f66-f3d1b80f3c7b" (UID: "07edb013-6500-498e-9f66-f3d1b80f3c7b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.774094 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07edb013-6500-498e-9f66-f3d1b80f3c7b-kube-api-access-h2lbr" (OuterVolumeSpecName: "kube-api-access-h2lbr") pod "07edb013-6500-498e-9f66-f3d1b80f3c7b" (UID: "07edb013-6500-498e-9f66-f3d1b80f3c7b"). InnerVolumeSpecName "kube-api-access-h2lbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.823764 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07edb013-6500-498e-9f66-f3d1b80f3c7b" (UID: "07edb013-6500-498e-9f66-f3d1b80f3c7b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.869207 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.869247 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2lbr\" (UniqueName: \"kubernetes.io/projected/07edb013-6500-498e-9f66-f3d1b80f3c7b-kube-api-access-h2lbr\") on node \"crc\" DevicePath \"\"" Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.869258 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07edb013-6500-498e-9f66-f3d1b80f3c7b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.997690 5025 generic.go:334] "Generic (PLEG): container finished" podID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerID="caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f" exitCode=0 Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.997737 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pz8s8" event={"ID":"07edb013-6500-498e-9f66-f3d1b80f3c7b","Type":"ContainerDied","Data":"caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f"} Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.997776 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pz8s8" event={"ID":"07edb013-6500-498e-9f66-f3d1b80f3c7b","Type":"ContainerDied","Data":"d4a05bada3ccfd3b1276dbeba5a17b45bc19ab488a71ae2aeed8af7fb356f599"} Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.997801 5025 scope.go:117] "RemoveContainer" containerID="caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f" Oct 04 10:48:44 crc kubenswrapper[5025]: I1004 10:48:44.997811 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pz8s8" Oct 04 10:48:45 crc kubenswrapper[5025]: I1004 10:48:45.042650 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pz8s8"] Oct 04 10:48:45 crc kubenswrapper[5025]: I1004 10:48:45.047311 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pz8s8"] Oct 04 10:48:45 crc kubenswrapper[5025]: I1004 10:48:45.289138 5025 scope.go:117] "RemoveContainer" containerID="176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88" Oct 04 10:48:45 crc kubenswrapper[5025]: I1004 10:48:45.317172 5025 scope.go:117] "RemoveContainer" containerID="a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0" Oct 04 10:48:45 crc kubenswrapper[5025]: I1004 10:48:45.353847 5025 scope.go:117] "RemoveContainer" containerID="caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f" Oct 04 10:48:45 crc kubenswrapper[5025]: E1004 10:48:45.354232 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f\": container with ID starting with caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f not found: ID does not exist" containerID="caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f" Oct 04 10:48:45 crc kubenswrapper[5025]: I1004 10:48:45.354266 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f"} err="failed to get container status \"caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f\": rpc error: code = NotFound desc = could not find container \"caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f\": container with ID starting with caa4ab38298dfcd489c2dbe3621b1e31edccba81a443997419571706dbfa768f not found: ID does not exist" Oct 04 10:48:45 crc kubenswrapper[5025]: I1004 10:48:45.354290 5025 scope.go:117] "RemoveContainer" containerID="176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88" Oct 04 10:48:45 crc kubenswrapper[5025]: E1004 10:48:45.354483 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88\": container with ID starting with 176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88 not found: ID does not exist" containerID="176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88" Oct 04 10:48:45 crc kubenswrapper[5025]: I1004 10:48:45.354512 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88"} err="failed to get container status \"176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88\": rpc error: code = NotFound desc = could not find container \"176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88\": container with ID starting with 176faddb069d4a1d98184ebacdd89869bf0aaa789e5a36aa589d6c47149a1a88 not found: ID does not exist" Oct 04 10:48:45 crc kubenswrapper[5025]: I1004 10:48:45.354528 5025 scope.go:117] "RemoveContainer" containerID="a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0" Oct 04 10:48:45 crc kubenswrapper[5025]: E1004 10:48:45.354715 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0\": container with ID starting with a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0 not found: ID does not exist" containerID="a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0" Oct 04 10:48:45 crc kubenswrapper[5025]: I1004 10:48:45.354742 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0"} err="failed to get container status \"a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0\": rpc error: code = NotFound desc = could not find container \"a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0\": container with ID starting with a09a32491d810d2af79c5d854a71b6d251c6b5bf00c1423a4eedd5d0b250ecf0 not found: ID does not exist" Oct 04 10:48:46 crc kubenswrapper[5025]: I1004 10:48:46.018034 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" event={"ID":"a1865149-283a-4933-b0d0-164afd0d1717","Type":"ContainerStarted","Data":"a06bf6b0193fb9c64396cd211fde7552c653e19bcd26de8c386ce397b501c989"} Oct 04 10:48:46 crc kubenswrapper[5025]: I1004 10:48:46.018437 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" Oct 04 10:48:46 crc kubenswrapper[5025]: I1004 10:48:46.066050 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" podStartSLOduration=1.589880041 podStartE2EDuration="8.066032698s" podCreationTimestamp="2025-10-04 10:48:38 +0000 UTC" firstStartedPulling="2025-10-04 10:48:38.853555893 +0000 UTC m=+847.278522773" lastFinishedPulling="2025-10-04 10:48:45.32970856 +0000 UTC m=+853.754675430" observedRunningTime="2025-10-04 10:48:46.063750962 +0000 UTC m=+854.488717862" watchObservedRunningTime="2025-10-04 10:48:46.066032698 +0000 UTC m=+854.490999578" Oct 04 10:48:46 crc kubenswrapper[5025]: I1004 10:48:46.424258 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07edb013-6500-498e-9f66-f3d1b80f3c7b" path="/var/lib/kubelet/pods/07edb013-6500-498e-9f66-f3d1b80f3c7b/volumes" Oct 04 10:48:47 crc kubenswrapper[5025]: I1004 10:48:47.027035 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-89c6d9c58-mtm9h" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.125957 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p"] Oct 04 10:49:12 crc kubenswrapper[5025]: E1004 10:49:12.126935 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerName="extract-content" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.126947 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerName="extract-content" Oct 04 10:49:12 crc kubenswrapper[5025]: E1004 10:49:12.126958 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerName="extract-utilities" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.126964 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerName="extract-utilities" Oct 04 10:49:12 crc kubenswrapper[5025]: E1004 10:49:12.126971 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerName="registry-server" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.126978 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerName="registry-server" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.127093 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="07edb013-6500-498e-9f66-f3d1b80f3c7b" containerName="registry-server" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.127716 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.131491 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-wtxn9" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.140221 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.142446 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.144197 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-njzwr" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.158122 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.163413 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.168336 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.169782 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.171990 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnrnd\" (UniqueName: \"kubernetes.io/projected/b2f1ef79-fb7a-4d17-ad22-08a2f46df01f-kube-api-access-rnrnd\") pod \"cinder-operator-controller-manager-7d4d4f8d-dn4sj\" (UID: \"b2f1ef79-fb7a-4d17-ad22-08a2f46df01f\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.172029 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ndl7\" (UniqueName: \"kubernetes.io/projected/91b206eb-240e-41b1-b1eb-df0c649554c9-kube-api-access-2ndl7\") pod \"barbican-operator-controller-manager-58c4cd55f4-mr64p\" (UID: \"91b206eb-240e-41b1-b1eb-df0c649554c9\") " pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.181470 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-vsvt6" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.190177 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.236568 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.237559 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.240037 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-dlxb6" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.245823 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.247268 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.255267 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-4mlmb" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.273077 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.273483 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m9qm\" (UniqueName: \"kubernetes.io/projected/b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228-kube-api-access-2m9qm\") pod \"glance-operator-controller-manager-5dc44df7d5-tx2zp\" (UID: \"b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228\") " pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.273527 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws2gh\" (UniqueName: \"kubernetes.io/projected/b2b93be0-ff60-4432-b0db-2ec0e6c605a2-kube-api-access-ws2gh\") pod \"heat-operator-controller-manager-54b4974c45-hqbgk\" (UID: \"b2b93be0-ff60-4432-b0db-2ec0e6c605a2\") " pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.273571 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pnx4\" (UniqueName: \"kubernetes.io/projected/0ed01e7a-3ef6-4cdd-97c5-ba85a84da593-kube-api-access-7pnx4\") pod \"designate-operator-controller-manager-75dfd9b554-ppf2w\" (UID: \"0ed01e7a-3ef6-4cdd-97c5-ba85a84da593\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.273609 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnrnd\" (UniqueName: \"kubernetes.io/projected/b2f1ef79-fb7a-4d17-ad22-08a2f46df01f-kube-api-access-rnrnd\") pod \"cinder-operator-controller-manager-7d4d4f8d-dn4sj\" (UID: \"b2f1ef79-fb7a-4d17-ad22-08a2f46df01f\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.273633 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ndl7\" (UniqueName: \"kubernetes.io/projected/91b206eb-240e-41b1-b1eb-df0c649554c9-kube-api-access-2ndl7\") pod \"barbican-operator-controller-manager-58c4cd55f4-mr64p\" (UID: \"91b206eb-240e-41b1-b1eb-df0c649554c9\") " pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.279770 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.280805 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.286318 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-k55s5" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.369664 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnrnd\" (UniqueName: \"kubernetes.io/projected/b2f1ef79-fb7a-4d17-ad22-08a2f46df01f-kube-api-access-rnrnd\") pod \"cinder-operator-controller-manager-7d4d4f8d-dn4sj\" (UID: \"b2f1ef79-fb7a-4d17-ad22-08a2f46df01f\") " pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.370595 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ndl7\" (UniqueName: \"kubernetes.io/projected/91b206eb-240e-41b1-b1eb-df0c649554c9-kube-api-access-2ndl7\") pod \"barbican-operator-controller-manager-58c4cd55f4-mr64p\" (UID: \"91b206eb-240e-41b1-b1eb-df0c649554c9\") " pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.399099 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.407214 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2dr9\" (UniqueName: \"kubernetes.io/projected/6f5b4526-b51d-46e6-900d-492cf48c2710-kube-api-access-n2dr9\") pod \"horizon-operator-controller-manager-76d5b87f47-f2blx\" (UID: \"6f5b4526-b51d-46e6-900d-492cf48c2710\") " pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.407265 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m9qm\" (UniqueName: \"kubernetes.io/projected/b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228-kube-api-access-2m9qm\") pod \"glance-operator-controller-manager-5dc44df7d5-tx2zp\" (UID: \"b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228\") " pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.407282 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws2gh\" (UniqueName: \"kubernetes.io/projected/b2b93be0-ff60-4432-b0db-2ec0e6c605a2-kube-api-access-ws2gh\") pod \"heat-operator-controller-manager-54b4974c45-hqbgk\" (UID: \"b2b93be0-ff60-4432-b0db-2ec0e6c605a2\") " pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.407337 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pnx4\" (UniqueName: \"kubernetes.io/projected/0ed01e7a-3ef6-4cdd-97c5-ba85a84da593-kube-api-access-7pnx4\") pod \"designate-operator-controller-manager-75dfd9b554-ppf2w\" (UID: \"0ed01e7a-3ef6-4cdd-97c5-ba85a84da593\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.411037 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.444686 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.446788 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws2gh\" (UniqueName: \"kubernetes.io/projected/b2b93be0-ff60-4432-b0db-2ec0e6c605a2-kube-api-access-ws2gh\") pod \"heat-operator-controller-manager-54b4974c45-hqbgk\" (UID: \"b2b93be0-ff60-4432-b0db-2ec0e6c605a2\") " pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.448372 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.449719 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.457611 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.457699 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pnx4\" (UniqueName: \"kubernetes.io/projected/0ed01e7a-3ef6-4cdd-97c5-ba85a84da593-kube-api-access-7pnx4\") pod \"designate-operator-controller-manager-75dfd9b554-ppf2w\" (UID: \"0ed01e7a-3ef6-4cdd-97c5-ba85a84da593\") " pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.461114 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.462399 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.476492 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-4bjb4" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.476869 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-9rzb6" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.477347 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m9qm\" (UniqueName: \"kubernetes.io/projected/b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228-kube-api-access-2m9qm\") pod \"glance-operator-controller-manager-5dc44df7d5-tx2zp\" (UID: \"b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228\") " pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.478791 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.487518 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.488752 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.491832 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.504331 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-nn9t9" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.511175 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.514171 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.515672 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2dr9\" (UniqueName: \"kubernetes.io/projected/6f5b4526-b51d-46e6-900d-492cf48c2710-kube-api-access-n2dr9\") pod \"horizon-operator-controller-manager-76d5b87f47-f2blx\" (UID: \"6f5b4526-b51d-46e6-900d-492cf48c2710\") " pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.515756 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdhcr\" (UniqueName: \"kubernetes.io/projected/36628a66-3fd5-4ca5-a6e4-f6d59009b69f-kube-api-access-mdhcr\") pod \"infra-operator-controller-manager-658588b8c9-mjrsp\" (UID: \"36628a66-3fd5-4ca5-a6e4-f6d59009b69f\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.515788 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrd8f\" (UniqueName: \"kubernetes.io/projected/8f554860-9fb5-4fc0-b795-832b03676469-kube-api-access-lrd8f\") pod \"ironic-operator-controller-manager-649675d675-f9rsb\" (UID: \"8f554860-9fb5-4fc0-b795-832b03676469\") " pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.515816 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hslf\" (UniqueName: \"kubernetes.io/projected/bd496b6b-d68a-466b-a7de-a11157a5f470-kube-api-access-5hslf\") pod \"keystone-operator-controller-manager-7b5ccf6d9c-ms9mz\" (UID: \"bd496b6b-d68a-466b-a7de-a11157a5f470\") " pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.515929 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36628a66-3fd5-4ca5-a6e4-f6d59009b69f-cert\") pod \"infra-operator-controller-manager-658588b8c9-mjrsp\" (UID: \"36628a66-3fd5-4ca5-a6e4-f6d59009b69f\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.524082 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.526269 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.527809 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.528832 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.532135 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.532251 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.537604 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.538930 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.539780 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.547290 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-s6dvm" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.547737 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7q7rf" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.547868 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.551409 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.555446 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.555629 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.561893 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.563963 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.567276 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.568185 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.569311 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.571640 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2dr9\" (UniqueName: \"kubernetes.io/projected/6f5b4526-b51d-46e6-900d-492cf48c2710-kube-api-access-n2dr9\") pod \"horizon-operator-controller-manager-76d5b87f47-f2blx\" (UID: \"6f5b4526-b51d-46e6-900d-492cf48c2710\") " pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.572439 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-bv87w" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.572591 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.573574 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.574591 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.575527 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.579423 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.582405 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.587173 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-drzp8" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.587724 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-pvj9c" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.587727 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.587916 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-f5ftk" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.588066 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.588946 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.592617 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.594326 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.599839 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-cpkzk" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.600235 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-nlpr2" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.608132 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-sjsp4" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.618326 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdhcr\" (UniqueName: \"kubernetes.io/projected/36628a66-3fd5-4ca5-a6e4-f6d59009b69f-kube-api-access-mdhcr\") pod \"infra-operator-controller-manager-658588b8c9-mjrsp\" (UID: \"36628a66-3fd5-4ca5-a6e4-f6d59009b69f\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.618363 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrd8f\" (UniqueName: \"kubernetes.io/projected/8f554860-9fb5-4fc0-b795-832b03676469-kube-api-access-lrd8f\") pod \"ironic-operator-controller-manager-649675d675-f9rsb\" (UID: \"8f554860-9fb5-4fc0-b795-832b03676469\") " pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.618388 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hslf\" (UniqueName: \"kubernetes.io/projected/bd496b6b-d68a-466b-a7de-a11157a5f470-kube-api-access-5hslf\") pod \"keystone-operator-controller-manager-7b5ccf6d9c-ms9mz\" (UID: \"bd496b6b-d68a-466b-a7de-a11157a5f470\") " pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.618463 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36628a66-3fd5-4ca5-a6e4-f6d59009b69f-cert\") pod \"infra-operator-controller-manager-658588b8c9-mjrsp\" (UID: \"36628a66-3fd5-4ca5-a6e4-f6d59009b69f\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:12 crc kubenswrapper[5025]: E1004 10:49:12.618611 5025 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 04 10:49:12 crc kubenswrapper[5025]: E1004 10:49:12.618660 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36628a66-3fd5-4ca5-a6e4-f6d59009b69f-cert podName:36628a66-3fd5-4ca5-a6e4-f6d59009b69f nodeName:}" failed. No retries permitted until 2025-10-04 10:49:13.118642717 +0000 UTC m=+881.543609597 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/36628a66-3fd5-4ca5-a6e4-f6d59009b69f-cert") pod "infra-operator-controller-manager-658588b8c9-mjrsp" (UID: "36628a66-3fd5-4ca5-a6e4-f6d59009b69f") : secret "infra-operator-webhook-server-cert" not found Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.626191 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.629858 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.639063 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.659170 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrd8f\" (UniqueName: \"kubernetes.io/projected/8f554860-9fb5-4fc0-b795-832b03676469-kube-api-access-lrd8f\") pod \"ironic-operator-controller-manager-649675d675-f9rsb\" (UID: \"8f554860-9fb5-4fc0-b795-832b03676469\") " pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.659708 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdhcr\" (UniqueName: \"kubernetes.io/projected/36628a66-3fd5-4ca5-a6e4-f6d59009b69f-kube-api-access-mdhcr\") pod \"infra-operator-controller-manager-658588b8c9-mjrsp\" (UID: \"36628a66-3fd5-4ca5-a6e4-f6d59009b69f\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.668658 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hslf\" (UniqueName: \"kubernetes.io/projected/bd496b6b-d68a-466b-a7de-a11157a5f470-kube-api-access-5hslf\") pod \"keystone-operator-controller-manager-7b5ccf6d9c-ms9mz\" (UID: \"bd496b6b-d68a-466b-a7de-a11157a5f470\") " pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.675635 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.676691 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.681380 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-8v84x" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.746445 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.750869 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e70c89d-b3a2-4b15-90eb-91449857b0a2-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx\" (UID: \"2e70c89d-b3a2-4b15-90eb-91449857b0a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.750978 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fm7b\" (UniqueName: \"kubernetes.io/projected/b8cb1e86-b4dd-48e9-85c2-6c688d283b7d-kube-api-access-2fm7b\") pod \"nova-operator-controller-manager-7c7fc454ff-4svdm\" (UID: \"b8cb1e86-b4dd-48e9-85c2-6c688d283b7d\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.751113 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr2st\" (UniqueName: \"kubernetes.io/projected/ad4743c5-3d51-421c-98b8-1463831fb92a-kube-api-access-lr2st\") pod \"swift-operator-controller-manager-6859f9b676-cl2kh\" (UID: \"ad4743c5-3d51-421c-98b8-1463831fb92a\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.751155 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6vtf\" (UniqueName: \"kubernetes.io/projected/2e70c89d-b3a2-4b15-90eb-91449857b0a2-kube-api-access-p6vtf\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx\" (UID: \"2e70c89d-b3a2-4b15-90eb-91449857b0a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.751193 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmb96\" (UniqueName: \"kubernetes.io/projected/81158795-65c7-4a01-b9ba-ad40b7d22582-kube-api-access-mmb96\") pod \"placement-operator-controller-manager-54689d9f88-f6bvd\" (UID: \"81158795-65c7-4a01-b9ba-ad40b7d22582\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.751356 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28flp\" (UniqueName: \"kubernetes.io/projected/dc8c4aca-63a0-490e-bf40-9377cdaefefc-kube-api-access-28flp\") pod \"neutron-operator-controller-manager-8d984cc4d-fbp6d\" (UID: \"dc8c4aca-63a0-490e-bf40-9377cdaefefc\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.751415 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wczx\" (UniqueName: \"kubernetes.io/projected/e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc-kube-api-access-7wczx\") pod \"octavia-operator-controller-manager-7468f855d8-2llfv\" (UID: \"e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.751652 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-289cs\" (UniqueName: \"kubernetes.io/projected/3ca07b7a-302b-474c-9aae-584e73350772-kube-api-access-289cs\") pod \"manila-operator-controller-manager-65d89cfd9f-r8csq\" (UID: \"3ca07b7a-302b-474c-9aae-584e73350772\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.751720 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nwqt\" (UniqueName: \"kubernetes.io/projected/b40c81d1-ba67-411d-b9ed-59a4991ac80f-kube-api-access-5nwqt\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx\" (UID: \"b40c81d1-ba67-411d-b9ed-59a4991ac80f\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.759990 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zkdm\" (UniqueName: \"kubernetes.io/projected/0371bb2b-35e9-4770-9fc4-560d5c2c343f-kube-api-access-2zkdm\") pod \"ovn-operator-controller-manager-6d8b6f9b9-dtnvw\" (UID: \"0371bb2b-35e9-4770-9fc4-560d5c2c343f\") " pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.843706 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.851049 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.874815 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.874901 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.880751 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-996zc" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884732 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr2st\" (UniqueName: \"kubernetes.io/projected/ad4743c5-3d51-421c-98b8-1463831fb92a-kube-api-access-lr2st\") pod \"swift-operator-controller-manager-6859f9b676-cl2kh\" (UID: \"ad4743c5-3d51-421c-98b8-1463831fb92a\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884770 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6vtf\" (UniqueName: \"kubernetes.io/projected/2e70c89d-b3a2-4b15-90eb-91449857b0a2-kube-api-access-p6vtf\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx\" (UID: \"2e70c89d-b3a2-4b15-90eb-91449857b0a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884798 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmb96\" (UniqueName: \"kubernetes.io/projected/81158795-65c7-4a01-b9ba-ad40b7d22582-kube-api-access-mmb96\") pod \"placement-operator-controller-manager-54689d9f88-f6bvd\" (UID: \"81158795-65c7-4a01-b9ba-ad40b7d22582\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884818 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28flp\" (UniqueName: \"kubernetes.io/projected/dc8c4aca-63a0-490e-bf40-9377cdaefefc-kube-api-access-28flp\") pod \"neutron-operator-controller-manager-8d984cc4d-fbp6d\" (UID: \"dc8c4aca-63a0-490e-bf40-9377cdaefefc\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884836 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wczx\" (UniqueName: \"kubernetes.io/projected/e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc-kube-api-access-7wczx\") pod \"octavia-operator-controller-manager-7468f855d8-2llfv\" (UID: \"e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884856 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdhjq\" (UniqueName: \"kubernetes.io/projected/85d359e6-3627-4c6b-b974-3599728017a3-kube-api-access-xdhjq\") pod \"telemetry-operator-controller-manager-5d4d74dd89-jrnhn\" (UID: \"85d359e6-3627-4c6b-b974-3599728017a3\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884901 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-289cs\" (UniqueName: \"kubernetes.io/projected/3ca07b7a-302b-474c-9aae-584e73350772-kube-api-access-289cs\") pod \"manila-operator-controller-manager-65d89cfd9f-r8csq\" (UID: \"3ca07b7a-302b-474c-9aae-584e73350772\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884935 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nwqt\" (UniqueName: \"kubernetes.io/projected/b40c81d1-ba67-411d-b9ed-59a4991ac80f-kube-api-access-5nwqt\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx\" (UID: \"b40c81d1-ba67-411d-b9ed-59a4991ac80f\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884959 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zkdm\" (UniqueName: \"kubernetes.io/projected/0371bb2b-35e9-4770-9fc4-560d5c2c343f-kube-api-access-2zkdm\") pod \"ovn-operator-controller-manager-6d8b6f9b9-dtnvw\" (UID: \"0371bb2b-35e9-4770-9fc4-560d5c2c343f\") " pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884981 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e70c89d-b3a2-4b15-90eb-91449857b0a2-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx\" (UID: \"2e70c89d-b3a2-4b15-90eb-91449857b0a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.884998 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fm7b\" (UniqueName: \"kubernetes.io/projected/b8cb1e86-b4dd-48e9-85c2-6c688d283b7d-kube-api-access-2fm7b\") pod \"nova-operator-controller-manager-7c7fc454ff-4svdm\" (UID: \"b8cb1e86-b4dd-48e9-85c2-6c688d283b7d\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" Oct 04 10:49:12 crc kubenswrapper[5025]: E1004 10:49:12.891400 5025 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 10:49:12 crc kubenswrapper[5025]: E1004 10:49:12.891490 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e70c89d-b3a2-4b15-90eb-91449857b0a2-cert podName:2e70c89d-b3a2-4b15-90eb-91449857b0a2 nodeName:}" failed. No retries permitted until 2025-10-04 10:49:13.391473646 +0000 UTC m=+881.816440526 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2e70c89d-b3a2-4b15-90eb-91449857b0a2-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" (UID: "2e70c89d-b3a2-4b15-90eb-91449857b0a2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.900040 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.900436 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.901437 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.914878 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl"] Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.927757 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-pvm26" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.936492 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wczx\" (UniqueName: \"kubernetes.io/projected/e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc-kube-api-access-7wczx\") pod \"octavia-operator-controller-manager-7468f855d8-2llfv\" (UID: \"e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc\") " pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.941990 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6vtf\" (UniqueName: \"kubernetes.io/projected/2e70c89d-b3a2-4b15-90eb-91449857b0a2-kube-api-access-p6vtf\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx\" (UID: \"2e70c89d-b3a2-4b15-90eb-91449857b0a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.943945 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28flp\" (UniqueName: \"kubernetes.io/projected/dc8c4aca-63a0-490e-bf40-9377cdaefefc-kube-api-access-28flp\") pod \"neutron-operator-controller-manager-8d984cc4d-fbp6d\" (UID: \"dc8c4aca-63a0-490e-bf40-9377cdaefefc\") " pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.947263 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zkdm\" (UniqueName: \"kubernetes.io/projected/0371bb2b-35e9-4770-9fc4-560d5c2c343f-kube-api-access-2zkdm\") pod \"ovn-operator-controller-manager-6d8b6f9b9-dtnvw\" (UID: \"0371bb2b-35e9-4770-9fc4-560d5c2c343f\") " pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.948724 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-289cs\" (UniqueName: \"kubernetes.io/projected/3ca07b7a-302b-474c-9aae-584e73350772-kube-api-access-289cs\") pod \"manila-operator-controller-manager-65d89cfd9f-r8csq\" (UID: \"3ca07b7a-302b-474c-9aae-584e73350772\") " pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.949189 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmb96\" (UniqueName: \"kubernetes.io/projected/81158795-65c7-4a01-b9ba-ad40b7d22582-kube-api-access-mmb96\") pod \"placement-operator-controller-manager-54689d9f88-f6bvd\" (UID: \"81158795-65c7-4a01-b9ba-ad40b7d22582\") " pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.961918 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr2st\" (UniqueName: \"kubernetes.io/projected/ad4743c5-3d51-421c-98b8-1463831fb92a-kube-api-access-lr2st\") pod \"swift-operator-controller-manager-6859f9b676-cl2kh\" (UID: \"ad4743c5-3d51-421c-98b8-1463831fb92a\") " pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.986522 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdhjq\" (UniqueName: \"kubernetes.io/projected/85d359e6-3627-4c6b-b974-3599728017a3-kube-api-access-xdhjq\") pod \"telemetry-operator-controller-manager-5d4d74dd89-jrnhn\" (UID: \"85d359e6-3627-4c6b-b974-3599728017a3\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.986695 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5js6\" (UniqueName: \"kubernetes.io/projected/a1cffbcd-c643-4b18-8832-d28767b38fc6-kube-api-access-j5js6\") pod \"test-operator-controller-manager-5cd5cb47d7-zmt6f\" (UID: \"a1cffbcd-c643-4b18-8832-d28767b38fc6\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.987058 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fm7b\" (UniqueName: \"kubernetes.io/projected/b8cb1e86-b4dd-48e9-85c2-6c688d283b7d-kube-api-access-2fm7b\") pod \"nova-operator-controller-manager-7c7fc454ff-4svdm\" (UID: \"b8cb1e86-b4dd-48e9-85c2-6c688d283b7d\") " pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.987538 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nwqt\" (UniqueName: \"kubernetes.io/projected/b40c81d1-ba67-411d-b9ed-59a4991ac80f-kube-api-access-5nwqt\") pod \"mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx\" (UID: \"b40c81d1-ba67-411d-b9ed-59a4991ac80f\") " pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" Oct 04 10:49:12 crc kubenswrapper[5025]: I1004 10:49:12.996774 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.031963 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t"] Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.033060 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.034616 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.034751 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-27jz2" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.043773 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t"] Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.047109 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdhjq\" (UniqueName: \"kubernetes.io/projected/85d359e6-3627-4c6b-b974-3599728017a3-kube-api-access-xdhjq\") pod \"telemetry-operator-controller-manager-5d4d74dd89-jrnhn\" (UID: \"85d359e6-3627-4c6b-b974-3599728017a3\") " pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.055989 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.061179 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml"] Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.062157 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml"] Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.062316 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.064500 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-wl5gr" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.070396 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.073502 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.088693 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5js6\" (UniqueName: \"kubernetes.io/projected/a1cffbcd-c643-4b18-8832-d28767b38fc6-kube-api-access-j5js6\") pod \"test-operator-controller-manager-5cd5cb47d7-zmt6f\" (UID: \"a1cffbcd-c643-4b18-8832-d28767b38fc6\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.088781 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzbrf\" (UniqueName: \"kubernetes.io/projected/80fb065c-ce02-4189-8b7a-7bbe78181ee9-kube-api-access-vzbrf\") pod \"watcher-operator-controller-manager-6cbc6dd547-lfgrl\" (UID: \"80fb065c-ce02-4189-8b7a-7bbe78181ee9\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.107990 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5js6\" (UniqueName: \"kubernetes.io/projected/a1cffbcd-c643-4b18-8832-d28767b38fc6-kube-api-access-j5js6\") pod \"test-operator-controller-manager-5cd5cb47d7-zmt6f\" (UID: \"a1cffbcd-c643-4b18-8832-d28767b38fc6\") " pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.148342 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.184400 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.195316 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/745ffa84-a65e-48f0-aced-495c5be76951-cert\") pod \"openstack-operator-controller-manager-8c8588487-9l88t\" (UID: \"745ffa84-a65e-48f0-aced-495c5be76951\") " pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.195388 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36628a66-3fd5-4ca5-a6e4-f6d59009b69f-cert\") pod \"infra-operator-controller-manager-658588b8c9-mjrsp\" (UID: \"36628a66-3fd5-4ca5-a6e4-f6d59009b69f\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.195445 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzbrf\" (UniqueName: \"kubernetes.io/projected/80fb065c-ce02-4189-8b7a-7bbe78181ee9-kube-api-access-vzbrf\") pod \"watcher-operator-controller-manager-6cbc6dd547-lfgrl\" (UID: \"80fb065c-ce02-4189-8b7a-7bbe78181ee9\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.195482 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlj7l\" (UniqueName: \"kubernetes.io/projected/a48ece71-fc88-4329-8ad3-fe9db58bb99a-kube-api-access-jlj7l\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml\" (UID: \"a48ece71-fc88-4329-8ad3-fe9db58bb99a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.195520 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9cbk\" (UniqueName: \"kubernetes.io/projected/745ffa84-a65e-48f0-aced-495c5be76951-kube-api-access-q9cbk\") pod \"openstack-operator-controller-manager-8c8588487-9l88t\" (UID: \"745ffa84-a65e-48f0-aced-495c5be76951\") " pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.214582 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.229623 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36628a66-3fd5-4ca5-a6e4-f6d59009b69f-cert\") pod \"infra-operator-controller-manager-658588b8c9-mjrsp\" (UID: \"36628a66-3fd5-4ca5-a6e4-f6d59009b69f\") " pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.230581 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzbrf\" (UniqueName: \"kubernetes.io/projected/80fb065c-ce02-4189-8b7a-7bbe78181ee9-kube-api-access-vzbrf\") pod \"watcher-operator-controller-manager-6cbc6dd547-lfgrl\" (UID: \"80fb065c-ce02-4189-8b7a-7bbe78181ee9\") " pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.249073 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.263349 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.279526 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.296886 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlj7l\" (UniqueName: \"kubernetes.io/projected/a48ece71-fc88-4329-8ad3-fe9db58bb99a-kube-api-access-jlj7l\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml\" (UID: \"a48ece71-fc88-4329-8ad3-fe9db58bb99a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.296951 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9cbk\" (UniqueName: \"kubernetes.io/projected/745ffa84-a65e-48f0-aced-495c5be76951-kube-api-access-q9cbk\") pod \"openstack-operator-controller-manager-8c8588487-9l88t\" (UID: \"745ffa84-a65e-48f0-aced-495c5be76951\") " pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.296996 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/745ffa84-a65e-48f0-aced-495c5be76951-cert\") pod \"openstack-operator-controller-manager-8c8588487-9l88t\" (UID: \"745ffa84-a65e-48f0-aced-495c5be76951\") " pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.298519 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.313499 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/745ffa84-a65e-48f0-aced-495c5be76951-cert\") pod \"openstack-operator-controller-manager-8c8588487-9l88t\" (UID: \"745ffa84-a65e-48f0-aced-495c5be76951\") " pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.315544 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlj7l\" (UniqueName: \"kubernetes.io/projected/a48ece71-fc88-4329-8ad3-fe9db58bb99a-kube-api-access-jlj7l\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml\" (UID: \"a48ece71-fc88-4329-8ad3-fe9db58bb99a\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.315556 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9cbk\" (UniqueName: \"kubernetes.io/projected/745ffa84-a65e-48f0-aced-495c5be76951-kube-api-access-q9cbk\") pod \"openstack-operator-controller-manager-8c8588487-9l88t\" (UID: \"745ffa84-a65e-48f0-aced-495c5be76951\") " pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.398509 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e70c89d-b3a2-4b15-90eb-91449857b0a2-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx\" (UID: \"2e70c89d-b3a2-4b15-90eb-91449857b0a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:13 crc kubenswrapper[5025]: E1004 10:49:13.398646 5025 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 10:49:13 crc kubenswrapper[5025]: E1004 10:49:13.398696 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e70c89d-b3a2-4b15-90eb-91449857b0a2-cert podName:2e70c89d-b3a2-4b15-90eb-91449857b0a2 nodeName:}" failed. No retries permitted until 2025-10-04 10:49:14.398682036 +0000 UTC m=+882.823648906 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2e70c89d-b3a2-4b15-90eb-91449857b0a2-cert") pod "openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" (UID: "2e70c89d-b3a2-4b15-90eb-91449857b0a2") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.412103 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.430462 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj"] Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.454026 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.475597 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.642455 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p"] Oct 04 10:49:13 crc kubenswrapper[5025]: W1004 10:49:13.658259 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91b206eb_240e_41b1_b1eb_df0c649554c9.slice/crio-2c0c544ef5cb4c9a6bbf5013a5fc3227a76d2a4a92f9b87364ec7f2b6f413a53 WatchSource:0}: Error finding container 2c0c544ef5cb4c9a6bbf5013a5fc3227a76d2a4a92f9b87364ec7f2b6f413a53: Status 404 returned error can't find the container with id 2c0c544ef5cb4c9a6bbf5013a5fc3227a76d2a4a92f9b87364ec7f2b6f413a53 Oct 04 10:49:13 crc kubenswrapper[5025]: I1004 10:49:13.898085 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.079943 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.097305 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.103552 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.111474 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.196293 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" event={"ID":"b2f1ef79-fb7a-4d17-ad22-08a2f46df01f","Type":"ContainerStarted","Data":"dbe6a225073862dffa7319a841903e0654763225d96c3046436587dcf2e1f1d1"} Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.197176 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" event={"ID":"8f554860-9fb5-4fc0-b795-832b03676469","Type":"ContainerStarted","Data":"cbd12fa588e2ee4154d06e70007cff3dd187dd72a69fff0681e69bff4a189b81"} Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.198521 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" event={"ID":"6f5b4526-b51d-46e6-900d-492cf48c2710","Type":"ContainerStarted","Data":"c2b915bd9963f9a82c8e54204a736a0958d6504a3e35dcca81d8369a8fae157c"} Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.199483 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" event={"ID":"91b206eb-240e-41b1-b1eb-df0c649554c9","Type":"ContainerStarted","Data":"2c0c544ef5cb4c9a6bbf5013a5fc3227a76d2a4a92f9b87364ec7f2b6f413a53"} Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.200196 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" event={"ID":"a1cffbcd-c643-4b18-8832-d28767b38fc6","Type":"ContainerStarted","Data":"11bbf81655b43614837c42caa7e4d45f95ea8a0fdd89a8bff8297402d988a3aa"} Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.201195 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" event={"ID":"dc8c4aca-63a0-490e-bf40-9377cdaefefc","Type":"ContainerStarted","Data":"bc16031774a3a50d19dad35e28c457b0e6f5dc6df0406720e87fde49e6441d80"} Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.202107 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" event={"ID":"3ca07b7a-302b-474c-9aae-584e73350772","Type":"ContainerStarted","Data":"724a61b173561dd8b1f47141c6480c7c3eab5663faa1815e90f9bb795e937e64"} Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.259123 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.266299 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.273390 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz"] Oct 04 10:49:14 crc kubenswrapper[5025]: W1004 10:49:14.284900 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2c63d15_fe9b_4c3f_bf1a_ac61c49fa228.slice/crio-8b26dfeb24fda143fbccbe063fe861ce881758a91ff0599f90d519026e5a3615 WatchSource:0}: Error finding container 8b26dfeb24fda143fbccbe063fe861ce881758a91ff0599f90d519026e5a3615: Status 404 returned error can't find the container with id 8b26dfeb24fda143fbccbe063fe861ce881758a91ff0599f90d519026e5a3615 Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.289999 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.419137 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e70c89d-b3a2-4b15-90eb-91449857b0a2-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx\" (UID: \"2e70c89d-b3a2-4b15-90eb-91449857b0a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.425894 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2e70c89d-b3a2-4b15-90eb-91449857b0a2-cert\") pod \"openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx\" (UID: \"2e70c89d-b3a2-4b15-90eb-91449857b0a2\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.489939 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.507563 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.519203 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.523779 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn"] Oct 04 10:49:14 crc kubenswrapper[5025]: W1004 10:49:14.524926 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad4743c5_3d51_421c_98b8_1463831fb92a.slice/crio-c607445f5788533298368ef2bcd7aff3d0cb0a34c6f5a8ea083cb847af16921e WatchSource:0}: Error finding container c607445f5788533298368ef2bcd7aff3d0cb0a34c6f5a8ea083cb847af16921e: Status 404 returned error can't find the container with id c607445f5788533298368ef2bcd7aff3d0cb0a34c6f5a8ea083cb847af16921e Oct 04 10:49:14 crc kubenswrapper[5025]: W1004 10:49:14.528992 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80fb065c_ce02_4189_8b7a_7bbe78181ee9.slice/crio-c95d9736cad662adca339df151a5dd90fdcdb331f7862d1b637db5aab9b2c6ff WatchSource:0}: Error finding container c95d9736cad662adca339df151a5dd90fdcdb331f7862d1b637db5aab9b2c6ff: Status 404 returned error can't find the container with id c95d9736cad662adca339df151a5dd90fdcdb331f7862d1b637db5aab9b2c6ff Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.539765 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.549638 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.553838 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv"] Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.559322 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mmb96,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-54689d9f88-f6bvd_openstack-operators(81158795-65c7-4a01-b9ba-ad40b7d22582): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.560792 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml"] Oct 04 10:49:14 crc kubenswrapper[5025]: W1004 10:49:14.565434 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode84cb66c_9fa4_4b67_bb7d_e70a2f06ccfc.slice/crio-15c53a1e9684dee181e9149936dce2853a5b28479088e8aad264fb0e364f122d WatchSource:0}: Error finding container 15c53a1e9684dee181e9149936dce2853a5b28479088e8aad264fb0e364f122d: Status 404 returned error can't find the container with id 15c53a1e9684dee181e9149936dce2853a5b28479088e8aad264fb0e364f122d Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.567927 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jlj7l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml_openstack-operators(a48ece71-fc88-4329-8ad3-fe9db58bb99a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.570867 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" podUID="a48ece71-fc88-4329-8ad3-fe9db58bb99a" Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.571040 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:5f96b563a63494082323bfced089d6589e0c89db43c6a39a2e912c79b1a278fe,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5nwqt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx_openstack-operators(b40c81d1-ba67-411d-b9ed-59a4991ac80f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.574269 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm"] Oct 04 10:49:14 crc kubenswrapper[5025]: W1004 10:49:14.579068 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8cb1e86_b4dd_48e9_85c2_6c688d283b7d.slice/crio-bd9fa3036758e12fc412be2937c088a1fdfd746cfc1d46794576c60b03a5cf69 WatchSource:0}: Error finding container bd9fa3036758e12fc412be2937c088a1fdfd746cfc1d46794576c60b03a5cf69: Status 404 returned error can't find the container with id bd9fa3036758e12fc412be2937c088a1fdfd746cfc1d46794576c60b03a5cf69 Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.579615 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw"] Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.585543 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t"] Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.594855 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2fm7b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-7c7fc454ff-4svdm_openstack-operators(b8cb1e86-b4dd-48e9-85c2-6c688d283b7d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.594871 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7wczx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7468f855d8-2llfv_openstack-operators(e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 10:49:14 crc kubenswrapper[5025]: W1004 10:49:14.604243 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0371bb2b_35e9_4770_9fc4_560d5c2c343f.slice/crio-f648fbc8b7af3bcbb9a80dbc10bf361c4e7a4e878bee86adc7fdb8b6115350f2 WatchSource:0}: Error finding container f648fbc8b7af3bcbb9a80dbc10bf361c4e7a4e878bee86adc7fdb8b6115350f2: Status 404 returned error can't find the container with id f648fbc8b7af3bcbb9a80dbc10bf361c4e7a4e878bee86adc7fdb8b6115350f2 Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.608296 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f37e29d1f621c23c0d77b09076006d1e8002a77c2ff3d9b8921f893221cb1d09,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2zkdm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-6d8b6f9b9-dtnvw_openstack-operators(0371bb2b-35e9-4770-9fc4-560d5c2c343f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 04 10:49:14 crc kubenswrapper[5025]: I1004 10:49:14.609609 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.762503 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" podUID="81158795-65c7-4a01-b9ba-ad40b7d22582" Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.775962 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" podUID="b40c81d1-ba67-411d-b9ed-59a4991ac80f" Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.836273 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" podUID="b8cb1e86-b4dd-48e9-85c2-6c688d283b7d" Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.844062 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" podUID="e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc" Oct 04 10:49:14 crc kubenswrapper[5025]: E1004 10:49:14.891362 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" podUID="0371bb2b-35e9-4770-9fc4-560d5c2c343f" Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.270796 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx"] Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.272816 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" event={"ID":"0ed01e7a-3ef6-4cdd-97c5-ba85a84da593","Type":"ContainerStarted","Data":"6a583d4f2bdb3eedd76f6d3bb2d32375bfa89c75fb582fabea147d2212dc2288"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.318218 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" event={"ID":"ad4743c5-3d51-421c-98b8-1463831fb92a","Type":"ContainerStarted","Data":"c607445f5788533298368ef2bcd7aff3d0cb0a34c6f5a8ea083cb847af16921e"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.348258 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" event={"ID":"a48ece71-fc88-4329-8ad3-fe9db58bb99a","Type":"ContainerStarted","Data":"9c1069371688f6f43c5919e43566b4c11b172781fae382f1bed956d317d17272"} Oct 04 10:49:15 crc kubenswrapper[5025]: E1004 10:49:15.356074 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" podUID="a48ece71-fc88-4329-8ad3-fe9db58bb99a" Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.378348 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" event={"ID":"80fb065c-ce02-4189-8b7a-7bbe78181ee9","Type":"ContainerStarted","Data":"c95d9736cad662adca339df151a5dd90fdcdb331f7862d1b637db5aab9b2c6ff"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.399814 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" event={"ID":"b40c81d1-ba67-411d-b9ed-59a4991ac80f","Type":"ContainerStarted","Data":"a6111708a53c14ff14ac6ef0cffc3df1c965d9cd86aba6890db322bb269910b1"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.399867 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" event={"ID":"b40c81d1-ba67-411d-b9ed-59a4991ac80f","Type":"ContainerStarted","Data":"3377258bf48f3310b3810efd40ded4a08791b2afb701cfa328a7fdd764870b83"} Oct 04 10:49:15 crc kubenswrapper[5025]: E1004 10:49:15.403130 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:5f96b563a63494082323bfced089d6589e0c89db43c6a39a2e912c79b1a278fe\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" podUID="b40c81d1-ba67-411d-b9ed-59a4991ac80f" Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.419817 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" event={"ID":"b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228","Type":"ContainerStarted","Data":"8b26dfeb24fda143fbccbe063fe861ce881758a91ff0599f90d519026e5a3615"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.433593 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" event={"ID":"745ffa84-a65e-48f0-aced-495c5be76951","Type":"ContainerStarted","Data":"7279863eddcddb1b5fa618b03ddaac9a01782f8eaebe2985d8b84a65118a7141"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.433633 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" event={"ID":"745ffa84-a65e-48f0-aced-495c5be76951","Type":"ContainerStarted","Data":"0bf454128536623fd66f60dfd9cad6970a59c0aacd2274e83206bd65e45eec41"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.433643 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" event={"ID":"745ffa84-a65e-48f0-aced-495c5be76951","Type":"ContainerStarted","Data":"ee6a123f0c2fb6ef6b51b8704222ff4bb8a828c0117671e74124b42abadcdf83"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.437451 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.444267 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" event={"ID":"b8cb1e86-b4dd-48e9-85c2-6c688d283b7d","Type":"ContainerStarted","Data":"bf6ee2ecbc8e5b9c7c1f2e316e958e91a34850366d777494d9810c388afba690"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.444302 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" event={"ID":"b8cb1e86-b4dd-48e9-85c2-6c688d283b7d","Type":"ContainerStarted","Data":"bd9fa3036758e12fc412be2937c088a1fdfd746cfc1d46794576c60b03a5cf69"} Oct 04 10:49:15 crc kubenswrapper[5025]: E1004 10:49:15.450197 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842\\\"\"" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" podUID="b8cb1e86-b4dd-48e9-85c2-6c688d283b7d" Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.455164 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" event={"ID":"bd496b6b-d68a-466b-a7de-a11157a5f470","Type":"ContainerStarted","Data":"210bcbf9b5070c11e6f7f608346e3f0818d0e3c3f47054351320cd3e431bf1a7"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.470631 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" podStartSLOduration=3.470614786 podStartE2EDuration="3.470614786s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:49:15.469436291 +0000 UTC m=+883.894403171" watchObservedRunningTime="2025-10-04 10:49:15.470614786 +0000 UTC m=+883.895581666" Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.472887 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" event={"ID":"b2b93be0-ff60-4432-b0db-2ec0e6c605a2","Type":"ContainerStarted","Data":"b0a02fa01f0574b60f7def2432d0759da1fea68925aa25f51ee6b9ceb09d65d2"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.478103 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" event={"ID":"81158795-65c7-4a01-b9ba-ad40b7d22582","Type":"ContainerStarted","Data":"410b80bd2e614fe8d09d436d15ffbf93b1b34c3f4db1495b9f784bc82ae5e217"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.478137 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" event={"ID":"81158795-65c7-4a01-b9ba-ad40b7d22582","Type":"ContainerStarted","Data":"d1f7b0fb0dadbf1642bc5b5de80e0e6b85c0ec829771569870bd576679286772"} Oct 04 10:49:15 crc kubenswrapper[5025]: E1004 10:49:15.479720 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" podUID="81158795-65c7-4a01-b9ba-ad40b7d22582" Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.480648 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" event={"ID":"85d359e6-3627-4c6b-b974-3599728017a3","Type":"ContainerStarted","Data":"32659d59c94f87ac89f8eb57e391c6a287a9b4d69c5b7d43ef43a716e8abe7b9"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.482330 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" event={"ID":"0371bb2b-35e9-4770-9fc4-560d5c2c343f","Type":"ContainerStarted","Data":"7ea8eb1e055af298e5c7bc08254fc3eeda414cf73767f1958133ef15fef5a4ce"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.482353 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" event={"ID":"0371bb2b-35e9-4770-9fc4-560d5c2c343f","Type":"ContainerStarted","Data":"f648fbc8b7af3bcbb9a80dbc10bf361c4e7a4e878bee86adc7fdb8b6115350f2"} Oct 04 10:49:15 crc kubenswrapper[5025]: E1004 10:49:15.489925 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f37e29d1f621c23c0d77b09076006d1e8002a77c2ff3d9b8921f893221cb1d09\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" podUID="0371bb2b-35e9-4770-9fc4-560d5c2c343f" Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.496690 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" event={"ID":"e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc","Type":"ContainerStarted","Data":"300477acb060041c9814e926e34da6a4e54e1d30bb24dbb09035fb14584b21e4"} Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.496739 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" event={"ID":"e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc","Type":"ContainerStarted","Data":"15c53a1e9684dee181e9149936dce2853a5b28479088e8aad264fb0e364f122d"} Oct 04 10:49:15 crc kubenswrapper[5025]: E1004 10:49:15.498952 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" podUID="e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc" Oct 04 10:49:15 crc kubenswrapper[5025]: I1004 10:49:15.508578 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" event={"ID":"36628a66-3fd5-4ca5-a6e4-f6d59009b69f","Type":"ContainerStarted","Data":"b3c5fcf486c17caa362055314bcd846f48ec826ced08778e4b32db69a8277e7a"} Oct 04 10:49:16 crc kubenswrapper[5025]: E1004 10:49:16.516751 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" podUID="81158795-65c7-4a01-b9ba-ad40b7d22582" Oct 04 10:49:16 crc kubenswrapper[5025]: E1004 10:49:16.516760 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f37e29d1f621c23c0d77b09076006d1e8002a77c2ff3d9b8921f893221cb1d09\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" podUID="0371bb2b-35e9-4770-9fc4-560d5c2c343f" Oct 04 10:49:16 crc kubenswrapper[5025]: E1004 10:49:16.516760 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:da5c3078d80878d66c616e6f8a0bb909f95d971cde2c612f96fded064113e182\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" podUID="e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc" Oct 04 10:49:16 crc kubenswrapper[5025]: E1004 10:49:16.516861 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" podUID="a48ece71-fc88-4329-8ad3-fe9db58bb99a" Oct 04 10:49:16 crc kubenswrapper[5025]: E1004 10:49:16.517474 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842\\\"\"" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" podUID="b8cb1e86-b4dd-48e9-85c2-6c688d283b7d" Oct 04 10:49:16 crc kubenswrapper[5025]: E1004 10:49:16.517482 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:5f96b563a63494082323bfced089d6589e0c89db43c6a39a2e912c79b1a278fe\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" podUID="b40c81d1-ba67-411d-b9ed-59a4991ac80f" Oct 04 10:49:17 crc kubenswrapper[5025]: W1004 10:49:17.558511 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e70c89d_b3a2_4b15_90eb_91449857b0a2.slice/crio-e650b52f03c71f443b21f09f26ac01742a596ede0599667b7182b862b236dcef WatchSource:0}: Error finding container e650b52f03c71f443b21f09f26ac01742a596ede0599667b7182b862b236dcef: Status 404 returned error can't find the container with id e650b52f03c71f443b21f09f26ac01742a596ede0599667b7182b862b236dcef Oct 04 10:49:18 crc kubenswrapper[5025]: I1004 10:49:18.530501 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" event={"ID":"2e70c89d-b3a2-4b15-90eb-91449857b0a2","Type":"ContainerStarted","Data":"e650b52f03c71f443b21f09f26ac01742a596ede0599667b7182b862b236dcef"} Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.419145 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-8c8588487-9l88t" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.515890 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b9gdk"] Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.518965 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.527852 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b9gdk"] Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.667862 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw5rs\" (UniqueName: \"kubernetes.io/projected/5c68f161-3c23-4b03-8789-961189e17d2d-kube-api-access-vw5rs\") pod \"certified-operators-b9gdk\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.667980 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-utilities\") pod \"certified-operators-b9gdk\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.668105 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-catalog-content\") pod \"certified-operators-b9gdk\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.769582 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw5rs\" (UniqueName: \"kubernetes.io/projected/5c68f161-3c23-4b03-8789-961189e17d2d-kube-api-access-vw5rs\") pod \"certified-operators-b9gdk\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.769656 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-utilities\") pod \"certified-operators-b9gdk\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.769719 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-catalog-content\") pod \"certified-operators-b9gdk\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.770277 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-catalog-content\") pod \"certified-operators-b9gdk\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.770349 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-utilities\") pod \"certified-operators-b9gdk\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.792742 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw5rs\" (UniqueName: \"kubernetes.io/projected/5c68f161-3c23-4b03-8789-961189e17d2d-kube-api-access-vw5rs\") pod \"certified-operators-b9gdk\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:23 crc kubenswrapper[5025]: I1004 10:49:23.855893 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:24 crc kubenswrapper[5025]: I1004 10:49:24.973121 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b9gdk"] Oct 04 10:49:24 crc kubenswrapper[5025]: W1004 10:49:24.989766 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5c68f161_3c23_4b03_8789_961189e17d2d.slice/crio-a7249e962f61b8a78e64c090470d03297ea2468395d57c6884e2b0998923b435 WatchSource:0}: Error finding container a7249e962f61b8a78e64c090470d03297ea2468395d57c6884e2b0998923b435: Status 404 returned error can't find the container with id a7249e962f61b8a78e64c090470d03297ea2468395d57c6884e2b0998923b435 Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.602354 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" event={"ID":"2e70c89d-b3a2-4b15-90eb-91449857b0a2","Type":"ContainerStarted","Data":"6926cd9eb17b938eddc1c50613173ff3cfc0c442c01f30410c278465b710315c"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.604639 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" event={"ID":"36628a66-3fd5-4ca5-a6e4-f6d59009b69f","Type":"ContainerStarted","Data":"02b340ae0f15bd0c6e40924957f535ba1764d1efdc28d19bcadc18d6e8a78297"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.619152 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" event={"ID":"3ca07b7a-302b-474c-9aae-584e73350772","Type":"ContainerStarted","Data":"7bbc2b48f9b0699b4497524300baa90353816621644a0843d1116d194dd962e6"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.619203 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.619214 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" event={"ID":"3ca07b7a-302b-474c-9aae-584e73350772","Type":"ContainerStarted","Data":"fdefaa3af187156a11f10ac1817f9ecb46db258d93b76044295961b976996b46"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.628540 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" event={"ID":"bd496b6b-d68a-466b-a7de-a11157a5f470","Type":"ContainerStarted","Data":"d153c0a7f8bd5e95fcbcbb7672209243126903c4300fa0a17f8c4aa594b39158"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.631996 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" event={"ID":"b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228","Type":"ContainerStarted","Data":"16a26cd216c05def8d6c996b6b1d5610acf55ae2ea52164fff76d8c2606b0980"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.632049 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" event={"ID":"b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228","Type":"ContainerStarted","Data":"d414961ecc46211f7c451ff0aa4669171848bed39b4962e40f84c7687ead06a8"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.632659 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.637276 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9gdk" event={"ID":"5c68f161-3c23-4b03-8789-961189e17d2d","Type":"ContainerStarted","Data":"a7249e962f61b8a78e64c090470d03297ea2468395d57c6884e2b0998923b435"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.639114 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" event={"ID":"8f554860-9fb5-4fc0-b795-832b03676469","Type":"ContainerStarted","Data":"905739896ad5499ced13cfaefba7c0187d93ac1aa3da4ede51ef50d4f397f66c"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.649384 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" event={"ID":"a1cffbcd-c643-4b18-8832-d28767b38fc6","Type":"ContainerStarted","Data":"f4d8934161149b9b3332d07a274a7e10776fe3bc36ec18d341922aaf62b66ec8"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.658041 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" podStartSLOduration=4.325025525 podStartE2EDuration="13.658005162s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.068901262 +0000 UTC m=+882.493868142" lastFinishedPulling="2025-10-04 10:49:23.401880899 +0000 UTC m=+891.826847779" observedRunningTime="2025-10-04 10:49:25.656854388 +0000 UTC m=+894.081821268" watchObservedRunningTime="2025-10-04 10:49:25.658005162 +0000 UTC m=+894.082972052" Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.663761 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" event={"ID":"dc8c4aca-63a0-490e-bf40-9377cdaefefc","Type":"ContainerStarted","Data":"45fd0f9c68861175f50d9b484758e25db1d0b0ef333134b08aa5ad40b476a793"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.663810 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" event={"ID":"dc8c4aca-63a0-490e-bf40-9377cdaefefc","Type":"ContainerStarted","Data":"4b1ec442425cb662df37a28994841bb86954a1ddd4a9ea2aa6eaeed8d1bfd5e1"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.664029 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.671201 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" event={"ID":"b2b93be0-ff60-4432-b0db-2ec0e6c605a2","Type":"ContainerStarted","Data":"f3b9a0adcc1e7b29fe4c87a2f13c00951277676ce7bd5162556c00d8162cb49f"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.677919 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" event={"ID":"80fb065c-ce02-4189-8b7a-7bbe78181ee9","Type":"ContainerStarted","Data":"46b7d602797a8c640c9322973bed11299603a0d3a2d36eccfc208a777999f14e"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.682835 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" event={"ID":"b2f1ef79-fb7a-4d17-ad22-08a2f46df01f","Type":"ContainerStarted","Data":"0830be0774938256e791ffc8b36e885bd96b0943d4fd879428c0c5b17ba1c29f"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.686957 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" event={"ID":"0ed01e7a-3ef6-4cdd-97c5-ba85a84da593","Type":"ContainerStarted","Data":"13c061562fa59905a15849962c9699592be4e5946b5e450369e0945bdbada4ca"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.687333 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" podStartSLOduration=3.48175774 podStartE2EDuration="13.687324614s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.288059841 +0000 UTC m=+882.713026721" lastFinishedPulling="2025-10-04 10:49:24.493626715 +0000 UTC m=+892.918593595" observedRunningTime="2025-10-04 10:49:25.686079408 +0000 UTC m=+894.111046288" watchObservedRunningTime="2025-10-04 10:49:25.687324614 +0000 UTC m=+894.112291494" Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.698516 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" event={"ID":"ad4743c5-3d51-421c-98b8-1463831fb92a","Type":"ContainerStarted","Data":"0fa99d20dc1f89670c00d1886a7f13d9bf32f01a1e07cec63f782616102db4cd"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.713575 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" event={"ID":"6f5b4526-b51d-46e6-900d-492cf48c2710","Type":"ContainerStarted","Data":"e768d7c0139bf80af50cb8c4fe9dc7c6790243781aad03876c90e9da93555b7d"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.713621 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" event={"ID":"6f5b4526-b51d-46e6-900d-492cf48c2710","Type":"ContainerStarted","Data":"1232b5073a41b504b1a82d2ab8e05e49eb31626ce109367e2d14aac8e6b34840"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.714180 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.714737 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" podStartSLOduration=3.302359447 podStartE2EDuration="13.71471889s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.080717256 +0000 UTC m=+882.505684136" lastFinishedPulling="2025-10-04 10:49:24.493076699 +0000 UTC m=+892.918043579" observedRunningTime="2025-10-04 10:49:25.714253446 +0000 UTC m=+894.139220326" watchObservedRunningTime="2025-10-04 10:49:25.71471889 +0000 UTC m=+894.139685770" Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.724289 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" event={"ID":"91b206eb-240e-41b1-b1eb-df0c649554c9","Type":"ContainerStarted","Data":"d93583cc70597e462f01f82f35247418c4816e2f35651fcacc1a6c2a044a5be6"} Oct 04 10:49:25 crc kubenswrapper[5025]: I1004 10:49:25.763679 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" podStartSLOduration=4.26497915 podStartE2EDuration="13.763664272s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:13.902035053 +0000 UTC m=+882.327001933" lastFinishedPulling="2025-10-04 10:49:23.400720175 +0000 UTC m=+891.825687055" observedRunningTime="2025-10-04 10:49:25.761351665 +0000 UTC m=+894.186318545" watchObservedRunningTime="2025-10-04 10:49:25.763664272 +0000 UTC m=+894.188631152" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.733507 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" event={"ID":"36628a66-3fd5-4ca5-a6e4-f6d59009b69f","Type":"ContainerStarted","Data":"aedeb5298cd93e4a2418a80d84fb7d1cea496846a2dae94ddb54cff1db005426"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.733915 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.735634 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" event={"ID":"91b206eb-240e-41b1-b1eb-df0c649554c9","Type":"ContainerStarted","Data":"18a2b6d1cc307c2d78761943b782560e9dab184885cab01c1fc0c5ebc3460045"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.735784 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.737206 5025 generic.go:334] "Generic (PLEG): container finished" podID="5c68f161-3c23-4b03-8789-961189e17d2d" containerID="b4c0069e8af371ad57376302efe02e0cb84f8b10ef6e16d2cc32a3c43d23417b" exitCode=0 Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.737279 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9gdk" event={"ID":"5c68f161-3c23-4b03-8789-961189e17d2d","Type":"ContainerDied","Data":"b4c0069e8af371ad57376302efe02e0cb84f8b10ef6e16d2cc32a3c43d23417b"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.738738 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" event={"ID":"8f554860-9fb5-4fc0-b795-832b03676469","Type":"ContainerStarted","Data":"36ac0bd4d827c7e3a2f47d37942d8e189976fd50d81f364e88259bf902cff74a"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.738853 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.741143 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" event={"ID":"2e70c89d-b3a2-4b15-90eb-91449857b0a2","Type":"ContainerStarted","Data":"901b0bccabc37a917f5730cb02be57215da2ee05e6b9041437a641647ec05ac9"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.741397 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.743686 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" event={"ID":"80fb065c-ce02-4189-8b7a-7bbe78181ee9","Type":"ContainerStarted","Data":"553e048715e3f428f662b2c9757b1caa6a53c1877102e8d69e61968cbcb8b2d6"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.744274 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.747308 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" event={"ID":"bd496b6b-d68a-466b-a7de-a11157a5f470","Type":"ContainerStarted","Data":"91ecae7bc68bbabe60012bdfdf39f2af12a281644e537dc174c6f3e3f01cc7b1"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.747710 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.749788 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" event={"ID":"b2f1ef79-fb7a-4d17-ad22-08a2f46df01f","Type":"ContainerStarted","Data":"8231ef4174aa7fff30c18a28dad23984979aa22f49793d9b5950542c23341e4c"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.749881 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.752040 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" event={"ID":"0ed01e7a-3ef6-4cdd-97c5-ba85a84da593","Type":"ContainerStarted","Data":"2ace1b7e3217562383a856b75d2792288e53e2d790f33d5a478730e53b795395"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.752156 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.753619 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" event={"ID":"ad4743c5-3d51-421c-98b8-1463831fb92a","Type":"ContainerStarted","Data":"e26f5601e5bc6628f509d9674568530456a53819ada00bc0adfa788628ee52cf"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.754496 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.758775 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" event={"ID":"85d359e6-3627-4c6b-b974-3599728017a3","Type":"ContainerStarted","Data":"18a74352b6e1211da1f46396a182c5a9c43c83b013b178a2cfb2b7e8e90cc450"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.758793 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" event={"ID":"85d359e6-3627-4c6b-b974-3599728017a3","Type":"ContainerStarted","Data":"7d0978ff7d3dd1ee1e11a06e2aec18a5e92096015e231bfe8f79d05d2c25dd89"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.759150 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.760228 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" event={"ID":"b2b93be0-ff60-4432-b0db-2ec0e6c605a2","Type":"ContainerStarted","Data":"37acf9ecf4f21e873e225ce4047175802e35c2feb3a6ea1b304b1a9dc2852b29"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.760567 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.762706 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" event={"ID":"a1cffbcd-c643-4b18-8832-d28767b38fc6","Type":"ContainerStarted","Data":"7faa326575cf052be04b919c1d920320659d38b9ca5042b2c1b0806fbb4700e4"} Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.762725 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.790650 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" podStartSLOduration=4.787837254 podStartE2EDuration="14.790634886s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.5275274 +0000 UTC m=+882.952494280" lastFinishedPulling="2025-10-04 10:49:24.530325042 +0000 UTC m=+892.955291912" observedRunningTime="2025-10-04 10:49:26.769393559 +0000 UTC m=+895.194360429" watchObservedRunningTime="2025-10-04 10:49:26.790634886 +0000 UTC m=+895.215601766" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.791177 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" podStartSLOduration=4.333263515 podStartE2EDuration="14.791173382s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.108285397 +0000 UTC m=+882.533252267" lastFinishedPulling="2025-10-04 10:49:24.566195254 +0000 UTC m=+892.991162134" observedRunningTime="2025-10-04 10:49:26.788314389 +0000 UTC m=+895.213281269" watchObservedRunningTime="2025-10-04 10:49:26.791173382 +0000 UTC m=+895.216140262" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.807467 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" podStartSLOduration=3.991629826 podStartE2EDuration="14.807455215s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:13.674224223 +0000 UTC m=+882.099191103" lastFinishedPulling="2025-10-04 10:49:24.490049622 +0000 UTC m=+892.915016492" observedRunningTime="2025-10-04 10:49:26.805482828 +0000 UTC m=+895.230449708" watchObservedRunningTime="2025-10-04 10:49:26.807455215 +0000 UTC m=+895.232422095" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.844297 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" podStartSLOduration=7.91089533 podStartE2EDuration="14.844280705s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:17.561182368 +0000 UTC m=+885.986149248" lastFinishedPulling="2025-10-04 10:49:24.494567743 +0000 UTC m=+892.919534623" observedRunningTime="2025-10-04 10:49:26.840730072 +0000 UTC m=+895.265696952" watchObservedRunningTime="2025-10-04 10:49:26.844280705 +0000 UTC m=+895.269247585" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.886145 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" podStartSLOduration=4.89538454 podStartE2EDuration="14.886126741s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.540799456 +0000 UTC m=+882.965766336" lastFinishedPulling="2025-10-04 10:49:24.531541657 +0000 UTC m=+892.956508537" observedRunningTime="2025-10-04 10:49:26.868332734 +0000 UTC m=+895.293299604" watchObservedRunningTime="2025-10-04 10:49:26.886126741 +0000 UTC m=+895.311093621" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.887636 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" podStartSLOduration=3.878269272 podStartE2EDuration="14.887629415s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:13.480432551 +0000 UTC m=+881.905399431" lastFinishedPulling="2025-10-04 10:49:24.489792694 +0000 UTC m=+892.914759574" observedRunningTime="2025-10-04 10:49:26.884292998 +0000 UTC m=+895.309259878" watchObservedRunningTime="2025-10-04 10:49:26.887629415 +0000 UTC m=+895.312596295" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.903639 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" podStartSLOduration=4.653543152 podStartE2EDuration="14.90362406s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.280256864 +0000 UTC m=+882.705223744" lastFinishedPulling="2025-10-04 10:49:24.530337772 +0000 UTC m=+892.955304652" observedRunningTime="2025-10-04 10:49:26.900645843 +0000 UTC m=+895.325612723" watchObservedRunningTime="2025-10-04 10:49:26.90362406 +0000 UTC m=+895.328590940" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.922187 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" podStartSLOduration=4.955305142 podStartE2EDuration="14.922165289s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.527200111 +0000 UTC m=+882.952166991" lastFinishedPulling="2025-10-04 10:49:24.494060258 +0000 UTC m=+892.919027138" observedRunningTime="2025-10-04 10:49:26.916758592 +0000 UTC m=+895.341725472" watchObservedRunningTime="2025-10-04 10:49:26.922165289 +0000 UTC m=+895.347132169" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.935102 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" podStartSLOduration=4.721982831 podStartE2EDuration="14.935081774s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.279971826 +0000 UTC m=+882.704938706" lastFinishedPulling="2025-10-04 10:49:24.493070769 +0000 UTC m=+892.918037649" observedRunningTime="2025-10-04 10:49:26.932601032 +0000 UTC m=+895.357567912" watchObservedRunningTime="2025-10-04 10:49:26.935081774 +0000 UTC m=+895.360048654" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.967698 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" podStartSLOduration=4.967621279 podStartE2EDuration="14.967676301s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.530844507 +0000 UTC m=+882.955811387" lastFinishedPulling="2025-10-04 10:49:24.530899529 +0000 UTC m=+892.955866409" observedRunningTime="2025-10-04 10:49:26.967069784 +0000 UTC m=+895.392036684" watchObservedRunningTime="2025-10-04 10:49:26.967676301 +0000 UTC m=+895.392643181" Oct 04 10:49:26 crc kubenswrapper[5025]: I1004 10:49:26.990081 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" podStartSLOduration=4.744631579 podStartE2EDuration="14.990058202s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.285144196 +0000 UTC m=+882.710111076" lastFinishedPulling="2025-10-04 10:49:24.530570819 +0000 UTC m=+892.955537699" observedRunningTime="2025-10-04 10:49:26.986595351 +0000 UTC m=+895.411562241" watchObservedRunningTime="2025-10-04 10:49:26.990058202 +0000 UTC m=+895.415025072" Oct 04 10:49:27 crc kubenswrapper[5025]: I1004 10:49:27.003949 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" podStartSLOduration=4.576007369 podStartE2EDuration="15.003934815s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.103719824 +0000 UTC m=+882.528686704" lastFinishedPulling="2025-10-04 10:49:24.53164727 +0000 UTC m=+892.956614150" observedRunningTime="2025-10-04 10:49:27.000073283 +0000 UTC m=+895.425040173" watchObservedRunningTime="2025-10-04 10:49:27.003934815 +0000 UTC m=+895.428901695" Oct 04 10:49:27 crc kubenswrapper[5025]: I1004 10:49:27.771231 5025 generic.go:334] "Generic (PLEG): container finished" podID="5c68f161-3c23-4b03-8789-961189e17d2d" containerID="c538a996e80530338cd68b735529fe99b30b5f8ee07b77945a217605fb6464a7" exitCode=0 Oct 04 10:49:27 crc kubenswrapper[5025]: I1004 10:49:27.771381 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9gdk" event={"ID":"5c68f161-3c23-4b03-8789-961189e17d2d","Type":"ContainerDied","Data":"c538a996e80530338cd68b735529fe99b30b5f8ee07b77945a217605fb6464a7"} Oct 04 10:49:28 crc kubenswrapper[5025]: I1004 10:49:28.782765 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9gdk" event={"ID":"5c68f161-3c23-4b03-8789-961189e17d2d","Type":"ContainerStarted","Data":"351a7ffbda4983239938285dc4fa38525c79eb6af8c6389eeb598da792932a1c"} Oct 04 10:49:28 crc kubenswrapper[5025]: I1004 10:49:28.807255 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b9gdk" podStartSLOduration=4.168915409 podStartE2EDuration="5.807225919s" podCreationTimestamp="2025-10-04 10:49:23 +0000 UTC" firstStartedPulling="2025-10-04 10:49:26.739419038 +0000 UTC m=+895.164385918" lastFinishedPulling="2025-10-04 10:49:28.377729548 +0000 UTC m=+896.802696428" observedRunningTime="2025-10-04 10:49:28.799126114 +0000 UTC m=+897.224093004" watchObservedRunningTime="2025-10-04 10:49:28.807225919 +0000 UTC m=+897.232192839" Oct 04 10:49:30 crc kubenswrapper[5025]: I1004 10:49:30.798097 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" event={"ID":"b40c81d1-ba67-411d-b9ed-59a4991ac80f","Type":"ContainerStarted","Data":"f95799e885ecbbe01cad777c0917a221d5ade1f948757cba8f7a900363dd82f7"} Oct 04 10:49:30 crc kubenswrapper[5025]: I1004 10:49:30.798588 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" Oct 04 10:49:30 crc kubenswrapper[5025]: I1004 10:49:30.799895 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" event={"ID":"e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc","Type":"ContainerStarted","Data":"925168f2939566ffec00396c966b9b6482b1b2923342bb3f628ebb1cda95da57"} Oct 04 10:49:30 crc kubenswrapper[5025]: I1004 10:49:30.800041 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" Oct 04 10:49:30 crc kubenswrapper[5025]: I1004 10:49:30.816992 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" podStartSLOduration=3.113063246 podStartE2EDuration="18.816970412s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.570934922 +0000 UTC m=+882.995901802" lastFinishedPulling="2025-10-04 10:49:30.274842088 +0000 UTC m=+898.699808968" observedRunningTime="2025-10-04 10:49:30.81484582 +0000 UTC m=+899.239812700" watchObservedRunningTime="2025-10-04 10:49:30.816970412 +0000 UTC m=+899.241937302" Oct 04 10:49:30 crc kubenswrapper[5025]: I1004 10:49:30.846991 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" podStartSLOduration=3.172119452 podStartE2EDuration="18.846968934s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.594764244 +0000 UTC m=+883.019731124" lastFinishedPulling="2025-10-04 10:49:30.269613726 +0000 UTC m=+898.694580606" observedRunningTime="2025-10-04 10:49:30.841256888 +0000 UTC m=+899.266223768" watchObservedRunningTime="2025-10-04 10:49:30.846968934 +0000 UTC m=+899.271935824" Oct 04 10:49:32 crc kubenswrapper[5025]: I1004 10:49:32.453391 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-58c4cd55f4-mr64p" Oct 04 10:49:32 crc kubenswrapper[5025]: I1004 10:49:32.460791 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7d4d4f8d-dn4sj" Oct 04 10:49:32 crc kubenswrapper[5025]: I1004 10:49:32.496376 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-75dfd9b554-ppf2w" Oct 04 10:49:32 crc kubenswrapper[5025]: I1004 10:49:32.590800 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5dc44df7d5-tx2zp" Oct 04 10:49:32 crc kubenswrapper[5025]: I1004 10:49:32.595439 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-54b4974c45-hqbgk" Oct 04 10:49:32 crc kubenswrapper[5025]: I1004 10:49:32.641983 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-76d5b87f47-f2blx" Oct 04 10:49:32 crc kubenswrapper[5025]: I1004 10:49:32.847971 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-649675d675-f9rsb" Oct 04 10:49:32 crc kubenswrapper[5025]: I1004 10:49:32.902755 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7b5ccf6d9c-ms9mz" Oct 04 10:49:33 crc kubenswrapper[5025]: I1004 10:49:33.000201 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-8d984cc4d-fbp6d" Oct 04 10:49:33 crc kubenswrapper[5025]: I1004 10:49:33.188521 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6859f9b676-cl2kh" Oct 04 10:49:33 crc kubenswrapper[5025]: I1004 10:49:33.227391 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5d4d74dd89-jrnhn" Oct 04 10:49:33 crc kubenswrapper[5025]: I1004 10:49:33.266314 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-65d89cfd9f-r8csq" Oct 04 10:49:33 crc kubenswrapper[5025]: I1004 10:49:33.268231 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cd5cb47d7-zmt6f" Oct 04 10:49:33 crc kubenswrapper[5025]: I1004 10:49:33.301990 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6cbc6dd547-lfgrl" Oct 04 10:49:33 crc kubenswrapper[5025]: I1004 10:49:33.485005 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-658588b8c9-mjrsp" Oct 04 10:49:33 crc kubenswrapper[5025]: I1004 10:49:33.856805 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:33 crc kubenswrapper[5025]: I1004 10:49:33.856945 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:33 crc kubenswrapper[5025]: I1004 10:49:33.895898 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:34 crc kubenswrapper[5025]: I1004 10:49:34.617712 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx" Oct 04 10:49:34 crc kubenswrapper[5025]: I1004 10:49:34.900385 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:35 crc kubenswrapper[5025]: I1004 10:49:35.049623 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b9gdk"] Oct 04 10:49:36 crc kubenswrapper[5025]: I1004 10:49:36.842507 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b9gdk" podUID="5c68f161-3c23-4b03-8789-961189e17d2d" containerName="registry-server" containerID="cri-o://351a7ffbda4983239938285dc4fa38525c79eb6af8c6389eeb598da792932a1c" gracePeriod=2 Oct 04 10:49:37 crc kubenswrapper[5025]: I1004 10:49:37.853756 5025 generic.go:334] "Generic (PLEG): container finished" podID="5c68f161-3c23-4b03-8789-961189e17d2d" containerID="351a7ffbda4983239938285dc4fa38525c79eb6af8c6389eeb598da792932a1c" exitCode=0 Oct 04 10:49:37 crc kubenswrapper[5025]: I1004 10:49:37.853803 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9gdk" event={"ID":"5c68f161-3c23-4b03-8789-961189e17d2d","Type":"ContainerDied","Data":"351a7ffbda4983239938285dc4fa38525c79eb6af8c6389eeb598da792932a1c"} Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.137801 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.196583 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw5rs\" (UniqueName: \"kubernetes.io/projected/5c68f161-3c23-4b03-8789-961189e17d2d-kube-api-access-vw5rs\") pod \"5c68f161-3c23-4b03-8789-961189e17d2d\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.196688 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-utilities\") pod \"5c68f161-3c23-4b03-8789-961189e17d2d\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.196726 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-catalog-content\") pod \"5c68f161-3c23-4b03-8789-961189e17d2d\" (UID: \"5c68f161-3c23-4b03-8789-961189e17d2d\") " Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.197471 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-utilities" (OuterVolumeSpecName: "utilities") pod "5c68f161-3c23-4b03-8789-961189e17d2d" (UID: "5c68f161-3c23-4b03-8789-961189e17d2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.202494 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c68f161-3c23-4b03-8789-961189e17d2d-kube-api-access-vw5rs" (OuterVolumeSpecName: "kube-api-access-vw5rs") pod "5c68f161-3c23-4b03-8789-961189e17d2d" (UID: "5c68f161-3c23-4b03-8789-961189e17d2d"). InnerVolumeSpecName "kube-api-access-vw5rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.266664 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c68f161-3c23-4b03-8789-961189e17d2d" (UID: "5c68f161-3c23-4b03-8789-961189e17d2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.298420 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw5rs\" (UniqueName: \"kubernetes.io/projected/5c68f161-3c23-4b03-8789-961189e17d2d-kube-api-access-vw5rs\") on node \"crc\" DevicePath \"\"" Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.298471 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.298490 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c68f161-3c23-4b03-8789-961189e17d2d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.881121 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b9gdk" event={"ID":"5c68f161-3c23-4b03-8789-961189e17d2d","Type":"ContainerDied","Data":"a7249e962f61b8a78e64c090470d03297ea2468395d57c6884e2b0998923b435"} Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.881189 5025 scope.go:117] "RemoveContainer" containerID="351a7ffbda4983239938285dc4fa38525c79eb6af8c6389eeb598da792932a1c" Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.881294 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b9gdk" Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.926959 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b9gdk"] Oct 04 10:49:39 crc kubenswrapper[5025]: I1004 10:49:39.931831 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b9gdk"] Oct 04 10:49:40 crc kubenswrapper[5025]: I1004 10:49:40.424098 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c68f161-3c23-4b03-8789-961189e17d2d" path="/var/lib/kubelet/pods/5c68f161-3c23-4b03-8789-961189e17d2d/volumes" Oct 04 10:49:42 crc kubenswrapper[5025]: I1004 10:49:42.783715 5025 scope.go:117] "RemoveContainer" containerID="c538a996e80530338cd68b735529fe99b30b5f8ee07b77945a217605fb6464a7" Oct 04 10:49:42 crc kubenswrapper[5025]: I1004 10:49:42.820948 5025 scope.go:117] "RemoveContainer" containerID="b4c0069e8af371ad57376302efe02e0cb84f8b10ef6e16d2cc32a3c43d23417b" Oct 04 10:49:43 crc kubenswrapper[5025]: I1004 10:49:43.075483 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7468f855d8-2llfv" Oct 04 10:49:43 crc kubenswrapper[5025]: I1004 10:49:43.284368 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx" Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.925065 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" event={"ID":"b8cb1e86-b4dd-48e9-85c2-6c688d283b7d","Type":"ContainerStarted","Data":"e8987636a48dbb7a1716d1f62780b14c61b47845a166dcc7c1a028da2b4350ec"} Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.925304 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.928534 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" event={"ID":"0371bb2b-35e9-4770-9fc4-560d5c2c343f","Type":"ContainerStarted","Data":"573774fcb0ded5ba7d2a3f76748cb1900c90273f5264396c2b1d15fbe422cf59"} Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.928764 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.930169 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" event={"ID":"a48ece71-fc88-4329-8ad3-fe9db58bb99a","Type":"ContainerStarted","Data":"b9da655fc98550477b3cb8aa011ee5de4ec27d0387de58f0a556d79fd0634d2e"} Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.932294 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" event={"ID":"81158795-65c7-4a01-b9ba-ad40b7d22582","Type":"ContainerStarted","Data":"c94380ac9811ab50e081a75c5c9585c7eb2ddd8676d076231183b72aeb134804"} Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.932519 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.944338 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" podStartSLOduration=4.754491825 podStartE2EDuration="32.944321311s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.594733353 +0000 UTC m=+883.019700233" lastFinishedPulling="2025-10-04 10:49:42.784562829 +0000 UTC m=+911.209529719" observedRunningTime="2025-10-04 10:49:44.942088596 +0000 UTC m=+913.367055476" watchObservedRunningTime="2025-10-04 10:49:44.944321311 +0000 UTC m=+913.369288191" Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.959030 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" podStartSLOduration=4.777482012 podStartE2EDuration="32.958992267s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.608195624 +0000 UTC m=+883.033162504" lastFinishedPulling="2025-10-04 10:49:42.789705859 +0000 UTC m=+911.214672759" observedRunningTime="2025-10-04 10:49:44.957470263 +0000 UTC m=+913.382437143" watchObservedRunningTime="2025-10-04 10:49:44.958992267 +0000 UTC m=+913.383959157" Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.978332 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" podStartSLOduration=4.750642314 podStartE2EDuration="32.978313089s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.55917169 +0000 UTC m=+882.984138570" lastFinishedPulling="2025-10-04 10:49:42.786842435 +0000 UTC m=+911.211809345" observedRunningTime="2025-10-04 10:49:44.9752675 +0000 UTC m=+913.400234390" watchObservedRunningTime="2025-10-04 10:49:44.978313089 +0000 UTC m=+913.403279969" Oct 04 10:49:44 crc kubenswrapper[5025]: I1004 10:49:44.997419 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml" podStartSLOduration=3.695926824 podStartE2EDuration="32.997400713s" podCreationTimestamp="2025-10-04 10:49:12 +0000 UTC" firstStartedPulling="2025-10-04 10:49:14.567812671 +0000 UTC m=+882.992779551" lastFinishedPulling="2025-10-04 10:49:43.86928657 +0000 UTC m=+912.294253440" observedRunningTime="2025-10-04 10:49:44.99281188 +0000 UTC m=+913.417778770" watchObservedRunningTime="2025-10-04 10:49:44.997400713 +0000 UTC m=+913.422367593" Oct 04 10:49:53 crc kubenswrapper[5025]: I1004 10:49:53.058915 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7c7fc454ff-4svdm" Oct 04 10:49:53 crc kubenswrapper[5025]: I1004 10:49:53.077223 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-6d8b6f9b9-dtnvw" Oct 04 10:49:53 crc kubenswrapper[5025]: I1004 10:49:53.151271 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-54689d9f88-f6bvd" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.852664 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7j2m"] Oct 04 10:50:08 crc kubenswrapper[5025]: E1004 10:50:08.853356 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c68f161-3c23-4b03-8789-961189e17d2d" containerName="extract-content" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.853368 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c68f161-3c23-4b03-8789-961189e17d2d" containerName="extract-content" Oct 04 10:50:08 crc kubenswrapper[5025]: E1004 10:50:08.853402 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c68f161-3c23-4b03-8789-961189e17d2d" containerName="registry-server" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.853408 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c68f161-3c23-4b03-8789-961189e17d2d" containerName="registry-server" Oct 04 10:50:08 crc kubenswrapper[5025]: E1004 10:50:08.853417 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c68f161-3c23-4b03-8789-961189e17d2d" containerName="extract-utilities" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.853424 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c68f161-3c23-4b03-8789-961189e17d2d" containerName="extract-utilities" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.853557 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c68f161-3c23-4b03-8789-961189e17d2d" containerName="registry-server" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.860362 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.863576 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-4szbx" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.863649 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.863852 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.863961 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.872603 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7j2m"] Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.956837 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5ztlr"] Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.958101 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.966413 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 04 10:50:08 crc kubenswrapper[5025]: I1004 10:50:08.974227 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5ztlr"] Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.057031 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j45d8\" (UniqueName: \"kubernetes.io/projected/74efd079-c7a5-44a1-a806-2c737aa0d6d9-kube-api-access-j45d8\") pod \"dnsmasq-dns-675f4bcbfc-h7j2m\" (UID: \"74efd079-c7a5-44a1-a806-2c737aa0d6d9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.057799 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74efd079-c7a5-44a1-a806-2c737aa0d6d9-config\") pod \"dnsmasq-dns-675f4bcbfc-h7j2m\" (UID: \"74efd079-c7a5-44a1-a806-2c737aa0d6d9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.159435 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-5ztlr\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.159511 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thq4l\" (UniqueName: \"kubernetes.io/projected/0069b904-3834-467e-bf04-1fd29f1045bb-kube-api-access-thq4l\") pod \"dnsmasq-dns-78dd6ddcc-5ztlr\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.159586 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74efd079-c7a5-44a1-a806-2c737aa0d6d9-config\") pod \"dnsmasq-dns-675f4bcbfc-h7j2m\" (UID: \"74efd079-c7a5-44a1-a806-2c737aa0d6d9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.159686 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j45d8\" (UniqueName: \"kubernetes.io/projected/74efd079-c7a5-44a1-a806-2c737aa0d6d9-kube-api-access-j45d8\") pod \"dnsmasq-dns-675f4bcbfc-h7j2m\" (UID: \"74efd079-c7a5-44a1-a806-2c737aa0d6d9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.159740 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-config\") pod \"dnsmasq-dns-78dd6ddcc-5ztlr\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.160542 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74efd079-c7a5-44a1-a806-2c737aa0d6d9-config\") pod \"dnsmasq-dns-675f4bcbfc-h7j2m\" (UID: \"74efd079-c7a5-44a1-a806-2c737aa0d6d9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.179530 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j45d8\" (UniqueName: \"kubernetes.io/projected/74efd079-c7a5-44a1-a806-2c737aa0d6d9-kube-api-access-j45d8\") pod \"dnsmasq-dns-675f4bcbfc-h7j2m\" (UID: \"74efd079-c7a5-44a1-a806-2c737aa0d6d9\") " pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.260559 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thq4l\" (UniqueName: \"kubernetes.io/projected/0069b904-3834-467e-bf04-1fd29f1045bb-kube-api-access-thq4l\") pod \"dnsmasq-dns-78dd6ddcc-5ztlr\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.260904 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-config\") pod \"dnsmasq-dns-78dd6ddcc-5ztlr\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.260946 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-5ztlr\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.261667 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-5ztlr\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.261783 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-config\") pod \"dnsmasq-dns-78dd6ddcc-5ztlr\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.276682 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thq4l\" (UniqueName: \"kubernetes.io/projected/0069b904-3834-467e-bf04-1fd29f1045bb-kube-api-access-thq4l\") pod \"dnsmasq-dns-78dd6ddcc-5ztlr\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.284707 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.478807 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.717732 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5ztlr"] Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.721322 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 10:50:09 crc kubenswrapper[5025]: I1004 10:50:09.948153 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7j2m"] Oct 04 10:50:10 crc kubenswrapper[5025]: I1004 10:50:10.127787 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" event={"ID":"74efd079-c7a5-44a1-a806-2c737aa0d6d9","Type":"ContainerStarted","Data":"234711fe66986fc80036f87f0fdf9366966c35c0357893248b738833f278acc3"} Oct 04 10:50:10 crc kubenswrapper[5025]: I1004 10:50:10.129485 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" event={"ID":"0069b904-3834-467e-bf04-1fd29f1045bb","Type":"ContainerStarted","Data":"d059b9a2d202cd0617a805de15b45a2e53d7afe2984ecd390912ff63ced33b1d"} Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.131732 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7j2m"] Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.137919 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7w6q7"] Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.139265 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.146889 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7w6q7"] Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.208405 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-config\") pod \"dnsmasq-dns-666b6646f7-7w6q7\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.208465 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp8pn\" (UniqueName: \"kubernetes.io/projected/8e141099-8d21-45e8-af84-e4595ddf8f1e-kube-api-access-cp8pn\") pod \"dnsmasq-dns-666b6646f7-7w6q7\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.208499 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7w6q7\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.309498 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-config\") pod \"dnsmasq-dns-666b6646f7-7w6q7\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.309555 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp8pn\" (UniqueName: \"kubernetes.io/projected/8e141099-8d21-45e8-af84-e4595ddf8f1e-kube-api-access-cp8pn\") pod \"dnsmasq-dns-666b6646f7-7w6q7\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.309663 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7w6q7\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.310426 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-config\") pod \"dnsmasq-dns-666b6646f7-7w6q7\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.310524 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-dns-svc\") pod \"dnsmasq-dns-666b6646f7-7w6q7\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.349720 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp8pn\" (UniqueName: \"kubernetes.io/projected/8e141099-8d21-45e8-af84-e4595ddf8f1e-kube-api-access-cp8pn\") pod \"dnsmasq-dns-666b6646f7-7w6q7\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.372856 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5ztlr"] Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.432325 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cszms"] Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.434300 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cszms"] Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.434382 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.466723 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.512625 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwwtm\" (UniqueName: \"kubernetes.io/projected/4ac1474e-ef0b-431e-9f1c-45053db21a37-kube-api-access-cwwtm\") pod \"dnsmasq-dns-57d769cc4f-cszms\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.512684 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-config\") pod \"dnsmasq-dns-57d769cc4f-cszms\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.512788 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cszms\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.614004 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwwtm\" (UniqueName: \"kubernetes.io/projected/4ac1474e-ef0b-431e-9f1c-45053db21a37-kube-api-access-cwwtm\") pod \"dnsmasq-dns-57d769cc4f-cszms\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.614051 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-config\") pod \"dnsmasq-dns-57d769cc4f-cszms\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.614129 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cszms\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.614870 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-cszms\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.615618 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-config\") pod \"dnsmasq-dns-57d769cc4f-cszms\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.642747 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwwtm\" (UniqueName: \"kubernetes.io/projected/4ac1474e-ef0b-431e-9f1c-45053db21a37-kube-api-access-cwwtm\") pod \"dnsmasq-dns-57d769cc4f-cszms\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.759089 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:12 crc kubenswrapper[5025]: W1004 10:50:12.980960 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e141099_8d21_45e8_af84_e4595ddf8f1e.slice/crio-fa22cd352bdd62db4857a29eaf0615b555caf40789932a2d41508b86d29e7894 WatchSource:0}: Error finding container fa22cd352bdd62db4857a29eaf0615b555caf40789932a2d41508b86d29e7894: Status 404 returned error can't find the container with id fa22cd352bdd62db4857a29eaf0615b555caf40789932a2d41508b86d29e7894 Oct 04 10:50:12 crc kubenswrapper[5025]: I1004 10:50:12.982889 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7w6q7"] Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.154175 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" event={"ID":"8e141099-8d21-45e8-af84-e4595ddf8f1e","Type":"ContainerStarted","Data":"fa22cd352bdd62db4857a29eaf0615b555caf40789932a2d41508b86d29e7894"} Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.195720 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cszms"] Oct 04 10:50:13 crc kubenswrapper[5025]: W1004 10:50:13.203102 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ac1474e_ef0b_431e_9f1c_45053db21a37.slice/crio-ac94eb095d8131a712e526bb1a237f5c8dd53fb5163ab1a4a38d3647b6b4b377 WatchSource:0}: Error finding container ac94eb095d8131a712e526bb1a237f5c8dd53fb5163ab1a4a38d3647b6b4b377: Status 404 returned error can't find the container with id ac94eb095d8131a712e526bb1a237f5c8dd53fb5163ab1a4a38d3647b6b4b377 Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.263432 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.265202 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.267602 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2kgnj" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.268638 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.271191 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.271396 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.271569 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.271768 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.272083 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.276908 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423477 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423528 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-876lf\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-kube-api-access-876lf\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423560 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-config-data\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423586 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423611 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423722 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423857 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423896 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423933 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a36a09c8-4254-4f14-bd39-7156cb462adb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423949 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a36a09c8-4254-4f14-bd39-7156cb462adb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.423967 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525380 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a36a09c8-4254-4f14-bd39-7156cb462adb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525425 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a36a09c8-4254-4f14-bd39-7156cb462adb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525457 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525520 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525542 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-876lf\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-kube-api-access-876lf\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525571 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-config-data\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525593 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525621 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525649 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525676 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.525719 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.526540 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.529670 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.530222 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-config-data\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.530400 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.531452 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.534447 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.538402 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a36a09c8-4254-4f14-bd39-7156cb462adb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.538697 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.538837 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.539869 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.540495 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a36a09c8-4254-4f14-bd39-7156cb462adb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.541282 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.544334 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.544557 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.544723 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.544817 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2vjs7" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.544954 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.545083 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.545347 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.548402 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.557212 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-876lf\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-kube-api-access-876lf\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.558106 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.593326 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.728659 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.728707 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.728726 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.728880 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.728934 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.729055 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfpqp\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-kube-api-access-kfpqp\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.729094 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.729118 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.729201 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.729269 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.729319 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.830701 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.830963 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831037 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831068 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831087 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831156 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831172 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831192 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfpqp\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-kube-api-access-kfpqp\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831210 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831229 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831257 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831472 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831544 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.831933 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.832278 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.833801 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.834299 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.836446 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.836566 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.837697 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.839714 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.849793 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfpqp\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-kube-api-access-kfpqp\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.858618 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:13 crc kubenswrapper[5025]: I1004 10:50:13.908228 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:50:14 crc kubenswrapper[5025]: I1004 10:50:14.012510 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 10:50:14 crc kubenswrapper[5025]: W1004 10:50:14.024414 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda36a09c8_4254_4f14_bd39_7156cb462adb.slice/crio-b46770e10f8fe09a9e912f0790a414fc2630132ceaca368b643521a8d28b2687 WatchSource:0}: Error finding container b46770e10f8fe09a9e912f0790a414fc2630132ceaca368b643521a8d28b2687: Status 404 returned error can't find the container with id b46770e10f8fe09a9e912f0790a414fc2630132ceaca368b643521a8d28b2687 Oct 04 10:50:14 crc kubenswrapper[5025]: I1004 10:50:14.161402 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a36a09c8-4254-4f14-bd39-7156cb462adb","Type":"ContainerStarted","Data":"b46770e10f8fe09a9e912f0790a414fc2630132ceaca368b643521a8d28b2687"} Oct 04 10:50:14 crc kubenswrapper[5025]: I1004 10:50:14.162683 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" event={"ID":"4ac1474e-ef0b-431e-9f1c-45053db21a37","Type":"ContainerStarted","Data":"ac94eb095d8131a712e526bb1a237f5c8dd53fb5163ab1a4a38d3647b6b4b377"} Oct 04 10:50:14 crc kubenswrapper[5025]: I1004 10:50:14.347101 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 10:50:15 crc kubenswrapper[5025]: I1004 10:50:15.987111 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 04 10:50:15 crc kubenswrapper[5025]: I1004 10:50:15.991342 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 10:50:15 crc kubenswrapper[5025]: I1004 10:50:15.996469 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 04 10:50:15 crc kubenswrapper[5025]: I1004 10:50:15.996732 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-zqjsx" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:15.998411 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:15.998585 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:15.998720 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.001135 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.007585 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.007848 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.010087 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.011804 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-9skql" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.012083 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.013749 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.014345 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.031614 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.165641 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.165676 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ee2b80-da98-4572-a2dc-c08ca7933e61-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.165703 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/64ee2b80-da98-4572-a2dc-c08ca7933e61-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.165772 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.165854 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/64ee2b80-da98-4572-a2dc-c08ca7933e61-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166001 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-secrets\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166149 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88xnn\" (UniqueName: \"kubernetes.io/projected/64ee2b80-da98-4572-a2dc-c08ca7933e61-kube-api-access-88xnn\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166205 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166253 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/64ee2b80-da98-4572-a2dc-c08ca7933e61-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166346 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166394 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-config-data-default\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166444 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166463 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64ee2b80-da98-4572-a2dc-c08ca7933e61-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166515 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h6s5\" (UniqueName: \"kubernetes.io/projected/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-kube-api-access-9h6s5\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166568 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166643 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ee2b80-da98-4572-a2dc-c08ca7933e61-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.166676 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/64ee2b80-da98-4572-a2dc-c08ca7933e61-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.167284 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-kolla-config\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.268832 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.268902 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/64ee2b80-da98-4572-a2dc-c08ca7933e61-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.268933 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.268964 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-config-data-default\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.268989 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.269454 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.269712 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/64ee2b80-da98-4572-a2dc-c08ca7933e61-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.270440 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-operator-scripts\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.272958 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-config-data-default\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.269006 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64ee2b80-da98-4572-a2dc-c08ca7933e61-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.274969 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h6s5\" (UniqueName: \"kubernetes.io/projected/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-kube-api-access-9h6s5\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.274998 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.275060 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ee2b80-da98-4572-a2dc-c08ca7933e61-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.275120 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/64ee2b80-da98-4572-a2dc-c08ca7933e61-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.276315 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64ee2b80-da98-4572-a2dc-c08ca7933e61-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.276836 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/64ee2b80-da98-4572-a2dc-c08ca7933e61-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.277540 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.277655 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-kolla-config\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.277834 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.277862 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ee2b80-da98-4572-a2dc-c08ca7933e61-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.277901 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/64ee2b80-da98-4572-a2dc-c08ca7933e61-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.277929 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.278579 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.279611 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/64ee2b80-da98-4572-a2dc-c08ca7933e61-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.280115 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/64ee2b80-da98-4572-a2dc-c08ca7933e61-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.280164 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.280394 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-config-data-generated\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.280584 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/64ee2b80-da98-4572-a2dc-c08ca7933e61-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.280651 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-secrets\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.280699 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88xnn\" (UniqueName: \"kubernetes.io/projected/64ee2b80-da98-4572-a2dc-c08ca7933e61-kube-api-access-88xnn\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.282318 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-kolla-config\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.283735 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/64ee2b80-da98-4572-a2dc-c08ca7933e61-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.295627 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.296234 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88xnn\" (UniqueName: \"kubernetes.io/projected/64ee2b80-da98-4572-a2dc-c08ca7933e61-kube-api-access-88xnn\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.297365 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h6s5\" (UniqueName: \"kubernetes.io/projected/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-kube-api-access-9h6s5\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.298969 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/64ee2b80-da98-4572-a2dc-c08ca7933e61-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"64ee2b80-da98-4572-a2dc-c08ca7933e61\") " pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.300967 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/5df194f4-9f9f-48e1-a4c9-87409b4f2b6e-secrets\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.304252 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e\") " pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.331257 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.344131 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.609189 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.610099 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.613432 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.614226 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.614443 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-gqn78" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.623480 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.792034 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2588531d-38e6-486e-b120-94bbd8a61b4f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.792093 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2588531d-38e6-486e-b120-94bbd8a61b4f-kolla-config\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.792123 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2588531d-38e6-486e-b120-94bbd8a61b4f-config-data\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.792215 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2588531d-38e6-486e-b120-94bbd8a61b4f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.792265 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5m8z\" (UniqueName: \"kubernetes.io/projected/2588531d-38e6-486e-b120-94bbd8a61b4f-kube-api-access-x5m8z\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.893128 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2588531d-38e6-486e-b120-94bbd8a61b4f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.893189 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2588531d-38e6-486e-b120-94bbd8a61b4f-kolla-config\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.893220 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2588531d-38e6-486e-b120-94bbd8a61b4f-config-data\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.893330 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2588531d-38e6-486e-b120-94bbd8a61b4f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.893384 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5m8z\" (UniqueName: \"kubernetes.io/projected/2588531d-38e6-486e-b120-94bbd8a61b4f-kube-api-access-x5m8z\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.893997 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2588531d-38e6-486e-b120-94bbd8a61b4f-kolla-config\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.894628 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2588531d-38e6-486e-b120-94bbd8a61b4f-config-data\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.896990 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2588531d-38e6-486e-b120-94bbd8a61b4f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.898958 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/2588531d-38e6-486e-b120-94bbd8a61b4f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.911600 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5m8z\" (UniqueName: \"kubernetes.io/projected/2588531d-38e6-486e-b120-94bbd8a61b4f-kube-api-access-x5m8z\") pod \"memcached-0\" (UID: \"2588531d-38e6-486e-b120-94bbd8a61b4f\") " pod="openstack/memcached-0" Oct 04 10:50:16 crc kubenswrapper[5025]: I1004 10:50:16.947285 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 04 10:50:18 crc kubenswrapper[5025]: I1004 10:50:18.642259 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 10:50:18 crc kubenswrapper[5025]: I1004 10:50:18.643443 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 10:50:18 crc kubenswrapper[5025]: I1004 10:50:18.646667 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-9zgr2" Oct 04 10:50:18 crc kubenswrapper[5025]: I1004 10:50:18.655820 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 10:50:18 crc kubenswrapper[5025]: I1004 10:50:18.734249 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxhzh\" (UniqueName: \"kubernetes.io/projected/8dbf11e2-e00c-4f47-b1cf-07223dac4842-kube-api-access-cxhzh\") pod \"kube-state-metrics-0\" (UID: \"8dbf11e2-e00c-4f47-b1cf-07223dac4842\") " pod="openstack/kube-state-metrics-0" Oct 04 10:50:18 crc kubenswrapper[5025]: I1004 10:50:18.835789 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxhzh\" (UniqueName: \"kubernetes.io/projected/8dbf11e2-e00c-4f47-b1cf-07223dac4842-kube-api-access-cxhzh\") pod \"kube-state-metrics-0\" (UID: \"8dbf11e2-e00c-4f47-b1cf-07223dac4842\") " pod="openstack/kube-state-metrics-0" Oct 04 10:50:18 crc kubenswrapper[5025]: I1004 10:50:18.866236 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxhzh\" (UniqueName: \"kubernetes.io/projected/8dbf11e2-e00c-4f47-b1cf-07223dac4842-kube-api-access-cxhzh\") pod \"kube-state-metrics-0\" (UID: \"8dbf11e2-e00c-4f47-b1cf-07223dac4842\") " pod="openstack/kube-state-metrics-0" Oct 04 10:50:18 crc kubenswrapper[5025]: I1004 10:50:18.963749 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 10:50:19 crc kubenswrapper[5025]: I1004 10:50:19.206713 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a","Type":"ContainerStarted","Data":"611be2be55fb857b34d266f1ee6ed53e3328c29a9e80563a5fd4dbcf3c7625b3"} Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.424638 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jcfxn"] Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.426123 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.432268 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-k5mm4" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.432363 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.435353 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.435588 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-c7fwx"] Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.437187 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.444289 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jcfxn"] Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.454085 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-c7fwx"] Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.542527 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.543783 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.546992 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.549726 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.549909 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.550076 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rhgxg" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.550231 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.550373 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.610684 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-var-log\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.610768 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-scripts\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.610811 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-var-run\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.610829 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-scripts\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.610848 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-etc-ovs\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.610888 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-var-run\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.610977 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-var-run-ovn\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.611036 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjbwp\" (UniqueName: \"kubernetes.io/projected/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-kube-api-access-pjbwp\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.611099 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-combined-ca-bundle\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.611142 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-var-lib\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.611198 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-var-log-ovn\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.611258 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh66f\" (UniqueName: \"kubernetes.io/projected/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-kube-api-access-rh66f\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.611304 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-ovn-controller-tls-certs\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.713828 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.713892 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-config\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.713936 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-var-log\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714032 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-scripts\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714082 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97ngl\" (UniqueName: \"kubernetes.io/projected/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-kube-api-access-97ngl\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714109 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-var-run\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714139 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-scripts\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714169 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-etc-ovs\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714189 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-var-run\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714229 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714269 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-var-run-ovn\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714302 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjbwp\" (UniqueName: \"kubernetes.io/projected/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-kube-api-access-pjbwp\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714326 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714374 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-combined-ca-bundle\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714421 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-var-lib\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714486 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-var-log-ovn\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714500 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-var-run\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714507 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714530 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714562 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714643 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-var-run\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714665 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh66f\" (UniqueName: \"kubernetes.io/projected/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-kube-api-access-rh66f\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714719 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-ovn-controller-tls-certs\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714959 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-var-run-ovn\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.715825 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-var-lib\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.715951 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-etc-ovs\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.715969 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-var-log-ovn\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.714561 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-var-log\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.717373 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-scripts\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.717979 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-scripts\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.732079 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-ovn-controller-tls-certs\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.732879 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-combined-ca-bundle\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.735598 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjbwp\" (UniqueName: \"kubernetes.io/projected/d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3-kube-api-access-pjbwp\") pod \"ovn-controller-ovs-c7fwx\" (UID: \"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3\") " pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.735849 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh66f\" (UniqueName: \"kubernetes.io/projected/fe21ee2a-4ed1-47aa-90f3-42629a279fd6-kube-api-access-rh66f\") pod \"ovn-controller-jcfxn\" (UID: \"fe21ee2a-4ed1-47aa-90f3-42629a279fd6\") " pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.754865 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.759988 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.815858 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97ngl\" (UniqueName: \"kubernetes.io/projected/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-kube-api-access-97ngl\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.815908 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.815944 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.816015 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.816098 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.816119 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.816166 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.816183 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-config\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.817104 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.817205 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-config\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.817356 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.817444 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.824899 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.825817 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.826578 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.839519 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97ngl\" (UniqueName: \"kubernetes.io/projected/c1026dd7-48ec-4cb2-aa1d-0e823fefea09-kube-api-access-97ngl\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.842535 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c1026dd7-48ec-4cb2-aa1d-0e823fefea09\") " pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:22 crc kubenswrapper[5025]: I1004 10:50:22.870514 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.528494 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.532877 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.537142 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.539821 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.540186 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.540340 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-xh55q" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.540493 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.657437 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9887ac7f-49c1-4316-92ba-5107f6f9260d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.657479 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9887ac7f-49c1-4316-92ba-5107f6f9260d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.657515 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9887ac7f-49c1-4316-92ba-5107f6f9260d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.657546 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9887ac7f-49c1-4316-92ba-5107f6f9260d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.657597 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9887ac7f-49c1-4316-92ba-5107f6f9260d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.657614 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.657886 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78hrq\" (UniqueName: \"kubernetes.io/projected/9887ac7f-49c1-4316-92ba-5107f6f9260d-kube-api-access-78hrq\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.658112 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9887ac7f-49c1-4316-92ba-5107f6f9260d-config\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.760108 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9887ac7f-49c1-4316-92ba-5107f6f9260d-config\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.760227 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9887ac7f-49c1-4316-92ba-5107f6f9260d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.760256 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9887ac7f-49c1-4316-92ba-5107f6f9260d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.760298 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9887ac7f-49c1-4316-92ba-5107f6f9260d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.760334 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9887ac7f-49c1-4316-92ba-5107f6f9260d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.760370 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9887ac7f-49c1-4316-92ba-5107f6f9260d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.760390 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.760433 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78hrq\" (UniqueName: \"kubernetes.io/projected/9887ac7f-49c1-4316-92ba-5107f6f9260d-kube-api-access-78hrq\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.760650 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9887ac7f-49c1-4316-92ba-5107f6f9260d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.760661 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.761359 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9887ac7f-49c1-4316-92ba-5107f6f9260d-config\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.761599 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9887ac7f-49c1-4316-92ba-5107f6f9260d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.766927 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9887ac7f-49c1-4316-92ba-5107f6f9260d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.768616 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9887ac7f-49c1-4316-92ba-5107f6f9260d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.769113 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9887ac7f-49c1-4316-92ba-5107f6f9260d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.776350 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78hrq\" (UniqueName: \"kubernetes.io/projected/9887ac7f-49c1-4316-92ba-5107f6f9260d-kube-api-access-78hrq\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.793042 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"9887ac7f-49c1-4316-92ba-5107f6f9260d\") " pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:25 crc kubenswrapper[5025]: I1004 10:50:25.862355 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:30 crc kubenswrapper[5025]: E1004 10:50:30.200993 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 10:50:30 crc kubenswrapper[5025]: E1004 10:50:30.201788 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-thq4l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-5ztlr_openstack(0069b904-3834-467e-bf04-1fd29f1045bb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 10:50:30 crc kubenswrapper[5025]: E1004 10:50:30.203007 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" podUID="0069b904-3834-467e-bf04-1fd29f1045bb" Oct 04 10:50:30 crc kubenswrapper[5025]: E1004 10:50:30.592097 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 10:50:30 crc kubenswrapper[5025]: E1004 10:50:30.592556 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j45d8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-h7j2m_openstack(74efd079-c7a5-44a1-a806-2c737aa0d6d9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 10:50:30 crc kubenswrapper[5025]: E1004 10:50:30.593749 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" podUID="74efd079-c7a5-44a1-a806-2c737aa0d6d9" Oct 04 10:50:30 crc kubenswrapper[5025]: I1004 10:50:30.669953 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 04 10:50:30 crc kubenswrapper[5025]: E1004 10:50:30.784321 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 04 10:50:30 crc kubenswrapper[5025]: E1004 10:50:30.784483 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cp8pn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-7w6q7_openstack(8e141099-8d21-45e8-af84-e4595ddf8f1e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 10:50:30 crc kubenswrapper[5025]: E1004 10:50:30.786175 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" podUID="8e141099-8d21-45e8-af84-e4595ddf8f1e" Oct 04 10:50:30 crc kubenswrapper[5025]: I1004 10:50:30.819139 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 10:50:30 crc kubenswrapper[5025]: I1004 10:50:30.828381 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 04 10:50:30 crc kubenswrapper[5025]: I1004 10:50:30.836905 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 04 10:50:30 crc kubenswrapper[5025]: I1004 10:50:30.969537 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 04 10:50:31 crc kubenswrapper[5025]: I1004 10:50:31.039283 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jcfxn"] Oct 04 10:50:31 crc kubenswrapper[5025]: I1004 10:50:31.110266 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 04 10:50:31 crc kubenswrapper[5025]: I1004 10:50:31.210293 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-c7fwx"] Oct 04 10:50:31 crc kubenswrapper[5025]: E1004 10:50:31.308967 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" podUID="8e141099-8d21-45e8-af84-e4595ddf8f1e" Oct 04 10:50:32 crc kubenswrapper[5025]: W1004 10:50:32.344398 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64ee2b80_da98_4572_a2dc_c08ca7933e61.slice/crio-0b43403309b4dd158d01b908b36d46a872dba6bf70ae764eb68a32096d93438c WatchSource:0}: Error finding container 0b43403309b4dd158d01b908b36d46a872dba6bf70ae764eb68a32096d93438c: Status 404 returned error can't find the container with id 0b43403309b4dd158d01b908b36d46a872dba6bf70ae764eb68a32096d93438c Oct 04 10:50:32 crc kubenswrapper[5025]: W1004 10:50:32.355871 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9887ac7f_49c1_4316_92ba_5107f6f9260d.slice/crio-48f3d12c5528dfdc83366747256056d84be347f5162c0ebb0227baeec8e70044 WatchSource:0}: Error finding container 48f3d12c5528dfdc83366747256056d84be347f5162c0ebb0227baeec8e70044: Status 404 returned error can't find the container with id 48f3d12c5528dfdc83366747256056d84be347f5162c0ebb0227baeec8e70044 Oct 04 10:50:32 crc kubenswrapper[5025]: W1004 10:50:32.359313 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8dbf11e2_e00c_4f47_b1cf_07223dac4842.slice/crio-a825b59431aeff480503a24f3513b7a36c75bb82611afaf11af14fba997a4dda WatchSource:0}: Error finding container a825b59431aeff480503a24f3513b7a36c75bb82611afaf11af14fba997a4dda: Status 404 returned error can't find the container with id a825b59431aeff480503a24f3513b7a36c75bb82611afaf11af14fba997a4dda Oct 04 10:50:32 crc kubenswrapper[5025]: W1004 10:50:32.382548 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1026dd7_48ec_4cb2_aa1d_0e823fefea09.slice/crio-ba76e68cd4b3ec803920eb71ac9beb2218d3db8a3297bf662c6feaa7b6ff7308 WatchSource:0}: Error finding container ba76e68cd4b3ec803920eb71ac9beb2218d3db8a3297bf662c6feaa7b6ff7308: Status 404 returned error can't find the container with id ba76e68cd4b3ec803920eb71ac9beb2218d3db8a3297bf662c6feaa7b6ff7308 Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.435947 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.441468 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.471981 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-config" (OuterVolumeSpecName: "config") pod "0069b904-3834-467e-bf04-1fd29f1045bb" (UID: "0069b904-3834-467e-bf04-1fd29f1045bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.472057 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-config\") pod \"0069b904-3834-467e-bf04-1fd29f1045bb\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.472171 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-dns-svc\") pod \"0069b904-3834-467e-bf04-1fd29f1045bb\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.472228 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thq4l\" (UniqueName: \"kubernetes.io/projected/0069b904-3834-467e-bf04-1fd29f1045bb-kube-api-access-thq4l\") pod \"0069b904-3834-467e-bf04-1fd29f1045bb\" (UID: \"0069b904-3834-467e-bf04-1fd29f1045bb\") " Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.472875 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.473368 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0069b904-3834-467e-bf04-1fd29f1045bb" (UID: "0069b904-3834-467e-bf04-1fd29f1045bb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.477557 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0069b904-3834-467e-bf04-1fd29f1045bb-kube-api-access-thq4l" (OuterVolumeSpecName: "kube-api-access-thq4l") pod "0069b904-3834-467e-bf04-1fd29f1045bb" (UID: "0069b904-3834-467e-bf04-1fd29f1045bb"). InnerVolumeSpecName "kube-api-access-thq4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.574415 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j45d8\" (UniqueName: \"kubernetes.io/projected/74efd079-c7a5-44a1-a806-2c737aa0d6d9-kube-api-access-j45d8\") pod \"74efd079-c7a5-44a1-a806-2c737aa0d6d9\" (UID: \"74efd079-c7a5-44a1-a806-2c737aa0d6d9\") " Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.574749 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74efd079-c7a5-44a1-a806-2c737aa0d6d9-config\") pod \"74efd079-c7a5-44a1-a806-2c737aa0d6d9\" (UID: \"74efd079-c7a5-44a1-a806-2c737aa0d6d9\") " Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.575176 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0069b904-3834-467e-bf04-1fd29f1045bb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.575196 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thq4l\" (UniqueName: \"kubernetes.io/projected/0069b904-3834-467e-bf04-1fd29f1045bb-kube-api-access-thq4l\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.575757 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74efd079-c7a5-44a1-a806-2c737aa0d6d9-config" (OuterVolumeSpecName: "config") pod "74efd079-c7a5-44a1-a806-2c737aa0d6d9" (UID: "74efd079-c7a5-44a1-a806-2c737aa0d6d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.582749 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74efd079-c7a5-44a1-a806-2c737aa0d6d9-kube-api-access-j45d8" (OuterVolumeSpecName: "kube-api-access-j45d8") pod "74efd079-c7a5-44a1-a806-2c737aa0d6d9" (UID: "74efd079-c7a5-44a1-a806-2c737aa0d6d9"). InnerVolumeSpecName "kube-api-access-j45d8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.676887 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j45d8\" (UniqueName: \"kubernetes.io/projected/74efd079-c7a5-44a1-a806-2c737aa0d6d9-kube-api-access-j45d8\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:32 crc kubenswrapper[5025]: I1004 10:50:32.676917 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74efd079-c7a5-44a1-a806-2c737aa0d6d9-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.328392 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" event={"ID":"0069b904-3834-467e-bf04-1fd29f1045bb","Type":"ContainerDied","Data":"d059b9a2d202cd0617a805de15b45a2e53d7afe2984ecd390912ff63ced33b1d"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.328691 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5ztlr" Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.330078 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2588531d-38e6-486e-b120-94bbd8a61b4f","Type":"ContainerStarted","Data":"b2ba102abbed2c8306c92195f336f5c9f2fe28ccc3c48734c9e6169e8fa9313a"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.331167 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c7fwx" event={"ID":"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3","Type":"ContainerStarted","Data":"dd8e6b01a32b0dbbb14a3037375cd5b6baf2884c43bcec40a27af4f5d38c9e90"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.332388 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"64ee2b80-da98-4572-a2dc-c08ca7933e61","Type":"ContainerStarted","Data":"0b43403309b4dd158d01b908b36d46a872dba6bf70ae764eb68a32096d93438c"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.334274 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9887ac7f-49c1-4316-92ba-5107f6f9260d","Type":"ContainerStarted","Data":"48f3d12c5528dfdc83366747256056d84be347f5162c0ebb0227baeec8e70044"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.335346 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" event={"ID":"74efd079-c7a5-44a1-a806-2c737aa0d6d9","Type":"ContainerDied","Data":"234711fe66986fc80036f87f0fdf9366966c35c0357893248b738833f278acc3"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.335512 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-h7j2m" Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.344130 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8dbf11e2-e00c-4f47-b1cf-07223dac4842","Type":"ContainerStarted","Data":"a825b59431aeff480503a24f3513b7a36c75bb82611afaf11af14fba997a4dda"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.345211 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jcfxn" event={"ID":"fe21ee2a-4ed1-47aa-90f3-42629a279fd6","Type":"ContainerStarted","Data":"a61cda7535d8a7d5b79fff9a1ef02488f9590d11f09214525030d5a7393f5ee6"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.350556 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" event={"ID":"4ac1474e-ef0b-431e-9f1c-45053db21a37","Type":"ContainerStarted","Data":"945b0c08823efba629a2ae16a6cf74170b83348a2d58213c51c7c829715f4c28"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.353138 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e","Type":"ContainerStarted","Data":"d22d4e37ab03de52ad99537a7dcc8c995b6e71e443f50ea38cc714c779144362"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.362578 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c1026dd7-48ec-4cb2-aa1d-0e823fefea09","Type":"ContainerStarted","Data":"ba76e68cd4b3ec803920eb71ac9beb2218d3db8a3297bf662c6feaa7b6ff7308"} Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.414129 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7j2m"] Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.418468 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-h7j2m"] Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.453927 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5ztlr"] Oct 04 10:50:33 crc kubenswrapper[5025]: I1004 10:50:33.470875 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5ztlr"] Oct 04 10:50:34 crc kubenswrapper[5025]: I1004 10:50:34.374470 5025 generic.go:334] "Generic (PLEG): container finished" podID="4ac1474e-ef0b-431e-9f1c-45053db21a37" containerID="945b0c08823efba629a2ae16a6cf74170b83348a2d58213c51c7c829715f4c28" exitCode=0 Oct 04 10:50:34 crc kubenswrapper[5025]: I1004 10:50:34.374549 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" event={"ID":"4ac1474e-ef0b-431e-9f1c-45053db21a37","Type":"ContainerDied","Data":"945b0c08823efba629a2ae16a6cf74170b83348a2d58213c51c7c829715f4c28"} Oct 04 10:50:34 crc kubenswrapper[5025]: I1004 10:50:34.421287 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0069b904-3834-467e-bf04-1fd29f1045bb" path="/var/lib/kubelet/pods/0069b904-3834-467e-bf04-1fd29f1045bb/volumes" Oct 04 10:50:34 crc kubenswrapper[5025]: I1004 10:50:34.421824 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74efd079-c7a5-44a1-a806-2c737aa0d6d9" path="/var/lib/kubelet/pods/74efd079-c7a5-44a1-a806-2c737aa0d6d9/volumes" Oct 04 10:50:35 crc kubenswrapper[5025]: I1004 10:50:35.385391 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" event={"ID":"4ac1474e-ef0b-431e-9f1c-45053db21a37","Type":"ContainerStarted","Data":"e7b213d0a5dcb4c83b558ae5cea062c2c698926a8444a969fa8f6e2b19d88b69"} Oct 04 10:50:35 crc kubenswrapper[5025]: I1004 10:50:35.385743 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:35 crc kubenswrapper[5025]: I1004 10:50:35.406714 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" podStartSLOduration=4.0292834299999996 podStartE2EDuration="23.406686218s" podCreationTimestamp="2025-10-04 10:50:12 +0000 UTC" firstStartedPulling="2025-10-04 10:50:13.205428489 +0000 UTC m=+941.630395369" lastFinishedPulling="2025-10-04 10:50:32.582831277 +0000 UTC m=+961.007798157" observedRunningTime="2025-10-04 10:50:35.40193837 +0000 UTC m=+963.826905260" watchObservedRunningTime="2025-10-04 10:50:35.406686218 +0000 UTC m=+963.831653098" Oct 04 10:50:36 crc kubenswrapper[5025]: I1004 10:50:36.396342 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a36a09c8-4254-4f14-bd39-7156cb462adb","Type":"ContainerStarted","Data":"66815974cd0d2555f9c56c6b7046d5ea1aa08f08cf16377d7c60422b077b160a"} Oct 04 10:50:36 crc kubenswrapper[5025]: I1004 10:50:36.403733 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a","Type":"ContainerStarted","Data":"2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea"} Oct 04 10:50:42 crc kubenswrapper[5025]: I1004 10:50:42.761197 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:42 crc kubenswrapper[5025]: I1004 10:50:42.805997 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7w6q7"] Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.187321 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.256275 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp8pn\" (UniqueName: \"kubernetes.io/projected/8e141099-8d21-45e8-af84-e4595ddf8f1e-kube-api-access-cp8pn\") pod \"8e141099-8d21-45e8-af84-e4595ddf8f1e\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.259630 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e141099-8d21-45e8-af84-e4595ddf8f1e-kube-api-access-cp8pn" (OuterVolumeSpecName: "kube-api-access-cp8pn") pod "8e141099-8d21-45e8-af84-e4595ddf8f1e" (UID: "8e141099-8d21-45e8-af84-e4595ddf8f1e"). InnerVolumeSpecName "kube-api-access-cp8pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.359802 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-config\") pod \"8e141099-8d21-45e8-af84-e4595ddf8f1e\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.359964 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-dns-svc\") pod \"8e141099-8d21-45e8-af84-e4595ddf8f1e\" (UID: \"8e141099-8d21-45e8-af84-e4595ddf8f1e\") " Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.360383 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-config" (OuterVolumeSpecName: "config") pod "8e141099-8d21-45e8-af84-e4595ddf8f1e" (UID: "8e141099-8d21-45e8-af84-e4595ddf8f1e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.360402 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp8pn\" (UniqueName: \"kubernetes.io/projected/8e141099-8d21-45e8-af84-e4595ddf8f1e-kube-api-access-cp8pn\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.360803 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e141099-8d21-45e8-af84-e4595ddf8f1e" (UID: "8e141099-8d21-45e8-af84-e4595ddf8f1e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.451404 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" event={"ID":"8e141099-8d21-45e8-af84-e4595ddf8f1e","Type":"ContainerDied","Data":"fa22cd352bdd62db4857a29eaf0615b555caf40789932a2d41508b86d29e7894"} Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.451457 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-7w6q7" Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.462393 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.462424 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e141099-8d21-45e8-af84-e4595ddf8f1e-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.515923 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7w6q7"] Oct 04 10:50:43 crc kubenswrapper[5025]: I1004 10:50:43.520717 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-7w6q7"] Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.420060 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e141099-8d21-45e8-af84-e4595ddf8f1e" path="/var/lib/kubelet/pods/8e141099-8d21-45e8-af84-e4595ddf8f1e/volumes" Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.466581 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c1026dd7-48ec-4cb2-aa1d-0e823fefea09","Type":"ContainerStarted","Data":"06052bbc30fb0cdf1c0250149da848b8ea3caee09678c152f51b63e1175d3511"} Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.468644 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"2588531d-38e6-486e-b120-94bbd8a61b4f","Type":"ContainerStarted","Data":"5ce45f280646c1318f94332f0ad15a6e3ff4a2fdea3e2a9e5bccb3ef6183dafd"} Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.468999 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.471853 5025 generic.go:334] "Generic (PLEG): container finished" podID="d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3" containerID="e85871d68b23349bc499e65d5cf321eb63692b0b9ed16779420eb64099a10c00" exitCode=0 Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.471929 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c7fwx" event={"ID":"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3","Type":"ContainerDied","Data":"e85871d68b23349bc499e65d5cf321eb63692b0b9ed16779420eb64099a10c00"} Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.474311 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e","Type":"ContainerStarted","Data":"a1cb76dbfd91f2818b4294ea4d2149287cb85390284229e91793c93dbd256b87"} Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.476784 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"64ee2b80-da98-4572-a2dc-c08ca7933e61","Type":"ContainerStarted","Data":"5d95d9cddbc8e1c8d32c7545e6d569af9ea585307118e7df2b99de6a0e52bbda"} Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.478589 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9887ac7f-49c1-4316-92ba-5107f6f9260d","Type":"ContainerStarted","Data":"ccddf109c21a84ac51f2c8984cccb38963c935b0f19864518f541e6e4d56a580"} Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.491562 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=18.668124676 podStartE2EDuration="28.491539548s" podCreationTimestamp="2025-10-04 10:50:16 +0000 UTC" firstStartedPulling="2025-10-04 10:50:32.531169875 +0000 UTC m=+960.956136755" lastFinishedPulling="2025-10-04 10:50:42.354584747 +0000 UTC m=+970.779551627" observedRunningTime="2025-10-04 10:50:44.483777932 +0000 UTC m=+972.908744812" watchObservedRunningTime="2025-10-04 10:50:44.491539548 +0000 UTC m=+972.916506428" Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.713471 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:50:44 crc kubenswrapper[5025]: I1004 10:50:44.713740 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.483899 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-84wmk"] Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.485608 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.488480 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.499198 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-84wmk"] Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.502901 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8dbf11e2-e00c-4f47-b1cf-07223dac4842","Type":"ContainerStarted","Data":"a0859cf37a7b14d6a0f6a57b9e16088a969791aa10112e37ce5663c76ab17ed5"} Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.503046 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.522544 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x27rx\" (UniqueName: \"kubernetes.io/projected/2954cd9e-67cf-4083-8ca7-d1e24449314b-kube-api-access-x27rx\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.522654 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2954cd9e-67cf-4083-8ca7-d1e24449314b-ovn-rundir\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.522679 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2954cd9e-67cf-4083-8ca7-d1e24449314b-config\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.522703 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2954cd9e-67cf-4083-8ca7-d1e24449314b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.522730 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2954cd9e-67cf-4083-8ca7-d1e24449314b-combined-ca-bundle\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.522752 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2954cd9e-67cf-4083-8ca7-d1e24449314b-ovs-rundir\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.523805 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jcfxn" event={"ID":"fe21ee2a-4ed1-47aa-90f3-42629a279fd6","Type":"ContainerStarted","Data":"414af1125381f4e0a77e6a7e0c84eb0417af6c0ca5ee4b7397ca2817cf863793"} Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.523946 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jcfxn" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.532880 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c7fwx" event={"ID":"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3","Type":"ContainerStarted","Data":"243b715c1ce1dfdbc44ab98be4112bfcdd8dd9ba4553574c5657aa92ceb03d7d"} Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.532932 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c7fwx" event={"ID":"d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3","Type":"ContainerStarted","Data":"d96774f3d28cfdac3c7c658d3e44374aeb7afb8f99054e314671f6c7429d8896"} Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.533098 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.533236 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.558363 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=15.808417172 podStartE2EDuration="27.558345516s" podCreationTimestamp="2025-10-04 10:50:18 +0000 UTC" firstStartedPulling="2025-10-04 10:50:32.531215717 +0000 UTC m=+960.956182607" lastFinishedPulling="2025-10-04 10:50:44.281144071 +0000 UTC m=+972.706110951" observedRunningTime="2025-10-04 10:50:45.528192284 +0000 UTC m=+973.953159164" watchObservedRunningTime="2025-10-04 10:50:45.558345516 +0000 UTC m=+973.983312396" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.585608 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jcfxn" podStartSLOduration=13.293581675 podStartE2EDuration="23.585587202s" podCreationTimestamp="2025-10-04 10:50:22 +0000 UTC" firstStartedPulling="2025-10-04 10:50:32.531542566 +0000 UTC m=+960.956509446" lastFinishedPulling="2025-10-04 10:50:42.823548093 +0000 UTC m=+971.248514973" observedRunningTime="2025-10-04 10:50:45.557052808 +0000 UTC m=+973.982019698" watchObservedRunningTime="2025-10-04 10:50:45.585587202 +0000 UTC m=+974.010554082" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.589874 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-c7fwx" podStartSLOduration=13.658563712 podStartE2EDuration="23.589849096s" podCreationTimestamp="2025-10-04 10:50:22 +0000 UTC" firstStartedPulling="2025-10-04 10:50:32.531215127 +0000 UTC m=+960.956182007" lastFinishedPulling="2025-10-04 10:50:42.462500511 +0000 UTC m=+970.887467391" observedRunningTime="2025-10-04 10:50:45.58484639 +0000 UTC m=+974.009813270" watchObservedRunningTime="2025-10-04 10:50:45.589849096 +0000 UTC m=+974.014815976" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.624646 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2954cd9e-67cf-4083-8ca7-d1e24449314b-ovn-rundir\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.624706 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2954cd9e-67cf-4083-8ca7-d1e24449314b-config\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.624736 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2954cd9e-67cf-4083-8ca7-d1e24449314b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.624771 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2954cd9e-67cf-4083-8ca7-d1e24449314b-combined-ca-bundle\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.624801 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2954cd9e-67cf-4083-8ca7-d1e24449314b-ovs-rundir\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.624856 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x27rx\" (UniqueName: \"kubernetes.io/projected/2954cd9e-67cf-4083-8ca7-d1e24449314b-kube-api-access-x27rx\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.625549 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2954cd9e-67cf-4083-8ca7-d1e24449314b-ovn-rundir\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.625664 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2954cd9e-67cf-4083-8ca7-d1e24449314b-ovs-rundir\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.625991 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2954cd9e-67cf-4083-8ca7-d1e24449314b-config\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.628293 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-n64qh"] Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.629797 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.637496 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2954cd9e-67cf-4083-8ca7-d1e24449314b-combined-ca-bundle\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.641953 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.646309 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-n64qh"] Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.649941 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2954cd9e-67cf-4083-8ca7-d1e24449314b-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.650056 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x27rx\" (UniqueName: \"kubernetes.io/projected/2954cd9e-67cf-4083-8ca7-d1e24449314b-kube-api-access-x27rx\") pod \"ovn-controller-metrics-84wmk\" (UID: \"2954cd9e-67cf-4083-8ca7-d1e24449314b\") " pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.814434 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-84wmk" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.835907 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.836123 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-config\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.836225 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.836253 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8984v\" (UniqueName: \"kubernetes.io/projected/ef3c7bdc-0f37-43fb-b247-3b043542f981-kube-api-access-8984v\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.892975 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-n64qh"] Oct 04 10:50:45 crc kubenswrapper[5025]: E1004 10:50:45.893506 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-8984v ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" podUID="ef3c7bdc-0f37-43fb-b247-3b043542f981" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.922317 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnsqw"] Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.923854 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.930351 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.938397 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.938479 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-config\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.938513 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.938534 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8984v\" (UniqueName: \"kubernetes.io/projected/ef3c7bdc-0f37-43fb-b247-3b043542f981-kube-api-access-8984v\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.939920 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.940470 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-config\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.940706 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.940714 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnsqw"] Oct 04 10:50:45 crc kubenswrapper[5025]: I1004 10:50:45.978044 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8984v\" (UniqueName: \"kubernetes.io/projected/ef3c7bdc-0f37-43fb-b247-3b043542f981-kube-api-access-8984v\") pod \"dnsmasq-dns-5bf47b49b7-n64qh\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.040399 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.044547 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.044611 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-dns-svc\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.044671 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-config\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.044841 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cn6t\" (UniqueName: \"kubernetes.io/projected/49e49e17-d84c-4614-8e56-02049a145e26-kube-api-access-5cn6t\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.146419 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-config\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.146524 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cn6t\" (UniqueName: \"kubernetes.io/projected/49e49e17-d84c-4614-8e56-02049a145e26-kube-api-access-5cn6t\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.146621 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.146682 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.146699 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-dns-svc\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.149474 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-dns-svc\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.150565 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.150706 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.152311 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-config\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.163702 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cn6t\" (UniqueName: \"kubernetes.io/projected/49e49e17-d84c-4614-8e56-02049a145e26-kube-api-access-5cn6t\") pod \"dnsmasq-dns-8554648995-mnsqw\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.279143 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.409000 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-84wmk"] Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.543078 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.543189 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-84wmk" event={"ID":"2954cd9e-67cf-4083-8ca7-d1e24449314b","Type":"ContainerStarted","Data":"181460fdd9aaafd0b75ff60555e6c44466ba14747da8cf5eb28412d4de4384e0"} Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.559412 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.653985 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-ovsdbserver-nb\") pod \"ef3c7bdc-0f37-43fb-b247-3b043542f981\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.654117 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-dns-svc\") pod \"ef3c7bdc-0f37-43fb-b247-3b043542f981\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.654292 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8984v\" (UniqueName: \"kubernetes.io/projected/ef3c7bdc-0f37-43fb-b247-3b043542f981-kube-api-access-8984v\") pod \"ef3c7bdc-0f37-43fb-b247-3b043542f981\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.654335 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-config\") pod \"ef3c7bdc-0f37-43fb-b247-3b043542f981\" (UID: \"ef3c7bdc-0f37-43fb-b247-3b043542f981\") " Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.654431 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ef3c7bdc-0f37-43fb-b247-3b043542f981" (UID: "ef3c7bdc-0f37-43fb-b247-3b043542f981"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.654580 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ef3c7bdc-0f37-43fb-b247-3b043542f981" (UID: "ef3c7bdc-0f37-43fb-b247-3b043542f981"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.654854 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-config" (OuterVolumeSpecName: "config") pod "ef3c7bdc-0f37-43fb-b247-3b043542f981" (UID: "ef3c7bdc-0f37-43fb-b247-3b043542f981"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.655043 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.655061 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.655074 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef3c7bdc-0f37-43fb-b247-3b043542f981-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.659769 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef3c7bdc-0f37-43fb-b247-3b043542f981-kube-api-access-8984v" (OuterVolumeSpecName: "kube-api-access-8984v") pod "ef3c7bdc-0f37-43fb-b247-3b043542f981" (UID: "ef3c7bdc-0f37-43fb-b247-3b043542f981"). InnerVolumeSpecName "kube-api-access-8984v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.731734 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnsqw"] Oct 04 10:50:46 crc kubenswrapper[5025]: I1004 10:50:46.756838 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8984v\" (UniqueName: \"kubernetes.io/projected/ef3c7bdc-0f37-43fb-b247-3b043542f981-kube-api-access-8984v\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:47 crc kubenswrapper[5025]: I1004 10:50:47.567635 5025 generic.go:334] "Generic (PLEG): container finished" podID="5df194f4-9f9f-48e1-a4c9-87409b4f2b6e" containerID="a1cb76dbfd91f2818b4294ea4d2149287cb85390284229e91793c93dbd256b87" exitCode=0 Oct 04 10:50:47 crc kubenswrapper[5025]: I1004 10:50:47.567707 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e","Type":"ContainerDied","Data":"a1cb76dbfd91f2818b4294ea4d2149287cb85390284229e91793c93dbd256b87"} Oct 04 10:50:47 crc kubenswrapper[5025]: I1004 10:50:47.568128 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-n64qh" Oct 04 10:50:47 crc kubenswrapper[5025]: I1004 10:50:47.628845 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-n64qh"] Oct 04 10:50:47 crc kubenswrapper[5025]: I1004 10:50:47.634688 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-n64qh"] Oct 04 10:50:47 crc kubenswrapper[5025]: W1004 10:50:47.838566 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49e49e17_d84c_4614_8e56_02049a145e26.slice/crio-dd887aaa060fbb1edbf182b5a38f6cd44c2bb26964691d6d9006a2145d862d69 WatchSource:0}: Error finding container dd887aaa060fbb1edbf182b5a38f6cd44c2bb26964691d6d9006a2145d862d69: Status 404 returned error can't find the container with id dd887aaa060fbb1edbf182b5a38f6cd44c2bb26964691d6d9006a2145d862d69 Oct 04 10:50:48 crc kubenswrapper[5025]: I1004 10:50:48.426904 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef3c7bdc-0f37-43fb-b247-3b043542f981" path="/var/lib/kubelet/pods/ef3c7bdc-0f37-43fb-b247-3b043542f981/volumes" Oct 04 10:50:48 crc kubenswrapper[5025]: I1004 10:50:48.575391 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"5df194f4-9f9f-48e1-a4c9-87409b4f2b6e","Type":"ContainerStarted","Data":"1873b3e2a6db9c623b171248d188b10aa5e17f8b0852e93daaf41e8b8c9f8ed9"} Oct 04 10:50:48 crc kubenswrapper[5025]: I1004 10:50:48.577346 5025 generic.go:334] "Generic (PLEG): container finished" podID="64ee2b80-da98-4572-a2dc-c08ca7933e61" containerID="5d95d9cddbc8e1c8d32c7545e6d569af9ea585307118e7df2b99de6a0e52bbda" exitCode=0 Oct 04 10:50:48 crc kubenswrapper[5025]: I1004 10:50:48.577407 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"64ee2b80-da98-4572-a2dc-c08ca7933e61","Type":"ContainerDied","Data":"5d95d9cddbc8e1c8d32c7545e6d569af9ea585307118e7df2b99de6a0e52bbda"} Oct 04 10:50:48 crc kubenswrapper[5025]: I1004 10:50:48.579557 5025 generic.go:334] "Generic (PLEG): container finished" podID="49e49e17-d84c-4614-8e56-02049a145e26" containerID="f4bd2a72b24d93cfa6908f1b9b80e79d7b0248f2e18f92c45a657e46a046d141" exitCode=0 Oct 04 10:50:48 crc kubenswrapper[5025]: I1004 10:50:48.579598 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-mnsqw" event={"ID":"49e49e17-d84c-4614-8e56-02049a145e26","Type":"ContainerDied","Data":"f4bd2a72b24d93cfa6908f1b9b80e79d7b0248f2e18f92c45a657e46a046d141"} Oct 04 10:50:48 crc kubenswrapper[5025]: I1004 10:50:48.579622 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-mnsqw" event={"ID":"49e49e17-d84c-4614-8e56-02049a145e26","Type":"ContainerStarted","Data":"dd887aaa060fbb1edbf182b5a38f6cd44c2bb26964691d6d9006a2145d862d69"} Oct 04 10:50:48 crc kubenswrapper[5025]: I1004 10:50:48.603453 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=24.187786737 podStartE2EDuration="34.603427387s" podCreationTimestamp="2025-10-04 10:50:14 +0000 UTC" firstStartedPulling="2025-10-04 10:50:32.531506995 +0000 UTC m=+960.956473875" lastFinishedPulling="2025-10-04 10:50:42.947147645 +0000 UTC m=+971.372114525" observedRunningTime="2025-10-04 10:50:48.593375633 +0000 UTC m=+977.018342523" watchObservedRunningTime="2025-10-04 10:50:48.603427387 +0000 UTC m=+977.028394277" Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.594863 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c1026dd7-48ec-4cb2-aa1d-0e823fefea09","Type":"ContainerStarted","Data":"6e3411f84466b0081fac2c7be3bddec5c5c2b046b318d1e18e655d81dbd624ad"} Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.598841 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"64ee2b80-da98-4572-a2dc-c08ca7933e61","Type":"ContainerStarted","Data":"00717a18b76d3efb463861f923c9ef3a83f85e522c72fd79363b12de60f7c63e"} Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.601582 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-84wmk" event={"ID":"2954cd9e-67cf-4083-8ca7-d1e24449314b","Type":"ContainerStarted","Data":"36d207039e8e9498b4ced8b591dced872c1ce09dce42632c6e1e87d7519b2b6d"} Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.606201 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"9887ac7f-49c1-4316-92ba-5107f6f9260d","Type":"ContainerStarted","Data":"29c73aabe0bb30989c40dc2504f2dc79ca2fa7edb8bba3dd0ad7a508cd4fc591"} Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.609740 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-mnsqw" event={"ID":"49e49e17-d84c-4614-8e56-02049a145e26","Type":"ContainerStarted","Data":"054a8e988af91704459b80f6746d1324e8ca08b74b076cb6e057906b86bf0906"} Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.610401 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.628820 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=12.75316784 podStartE2EDuration="28.628786354s" podCreationTimestamp="2025-10-04 10:50:21 +0000 UTC" firstStartedPulling="2025-10-04 10:50:32.531216547 +0000 UTC m=+960.956183417" lastFinishedPulling="2025-10-04 10:50:48.406835051 +0000 UTC m=+976.831801931" observedRunningTime="2025-10-04 10:50:49.622779798 +0000 UTC m=+978.047746708" watchObservedRunningTime="2025-10-04 10:50:49.628786354 +0000 UTC m=+978.053753284" Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.678618 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-mnsqw" podStartSLOduration=4.678587219 podStartE2EDuration="4.678587219s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:50:49.673922653 +0000 UTC m=+978.098889603" watchObservedRunningTime="2025-10-04 10:50:49.678587219 +0000 UTC m=+978.103554139" Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.705734 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=9.848593387 podStartE2EDuration="25.705678231s" podCreationTimestamp="2025-10-04 10:50:24 +0000 UTC" firstStartedPulling="2025-10-04 10:50:32.531245817 +0000 UTC m=+960.956212717" lastFinishedPulling="2025-10-04 10:50:48.388330681 +0000 UTC m=+976.813297561" observedRunningTime="2025-10-04 10:50:49.70360426 +0000 UTC m=+978.128571160" watchObservedRunningTime="2025-10-04 10:50:49.705678231 +0000 UTC m=+978.130645121" Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.740672 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=25.344664888 podStartE2EDuration="35.740652633s" podCreationTimestamp="2025-10-04 10:50:14 +0000 UTC" firstStartedPulling="2025-10-04 10:50:32.531254998 +0000 UTC m=+960.956221878" lastFinishedPulling="2025-10-04 10:50:42.927242743 +0000 UTC m=+971.352209623" observedRunningTime="2025-10-04 10:50:49.737934683 +0000 UTC m=+978.162901583" watchObservedRunningTime="2025-10-04 10:50:49.740652633 +0000 UTC m=+978.165619523" Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.759216 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-84wmk" podStartSLOduration=2.809334732 podStartE2EDuration="4.759196135s" podCreationTimestamp="2025-10-04 10:50:45 +0000 UTC" firstStartedPulling="2025-10-04 10:50:46.439075835 +0000 UTC m=+974.864042715" lastFinishedPulling="2025-10-04 10:50:48.388937238 +0000 UTC m=+976.813904118" observedRunningTime="2025-10-04 10:50:49.758449803 +0000 UTC m=+978.183416693" watchObservedRunningTime="2025-10-04 10:50:49.759196135 +0000 UTC m=+978.184163025" Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.863176 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.872115 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.926157 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:49 crc kubenswrapper[5025]: I1004 10:50:49.935676 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:50 crc kubenswrapper[5025]: I1004 10:50:50.617859 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:50 crc kubenswrapper[5025]: I1004 10:50:50.618980 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:50 crc kubenswrapper[5025]: I1004 10:50:50.677307 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 04 10:50:50 crc kubenswrapper[5025]: I1004 10:50:50.688960 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.099918 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.101156 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.102539 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.102908 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.103175 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-z5kjr" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.112592 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.143157 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.237470 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.237552 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t25mq\" (UniqueName: \"kubernetes.io/projected/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-kube-api-access-t25mq\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.237707 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-config\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.237805 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-scripts\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.237855 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.237973 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.238096 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.339319 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t25mq\" (UniqueName: \"kubernetes.io/projected/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-kube-api-access-t25mq\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.339369 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-config\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.339406 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-scripts\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.339455 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.339485 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.340275 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.340468 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-scripts\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.340559 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-config\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.340967 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.341098 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.345810 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.345855 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.350827 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.363923 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t25mq\" (UniqueName: \"kubernetes.io/projected/ee98710d-b3c1-4dbf-ab50-ee831c4d622d-kube-api-access-t25mq\") pod \"ovn-northd-0\" (UID: \"ee98710d-b3c1-4dbf-ab50-ee831c4d622d\") " pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.464286 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.950231 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 04 10:50:51 crc kubenswrapper[5025]: I1004 10:50:51.951340 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 04 10:50:51 crc kubenswrapper[5025]: W1004 10:50:51.964332 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee98710d_b3c1_4dbf_ab50_ee831c4d622d.slice/crio-704061e875bd0d83310d6943a8f92ab71e004a1c5c9f89f26f568828f02239fe WatchSource:0}: Error finding container 704061e875bd0d83310d6943a8f92ab71e004a1c5c9f89f26f568828f02239fe: Status 404 returned error can't find the container with id 704061e875bd0d83310d6943a8f92ab71e004a1c5c9f89f26f568828f02239fe Oct 04 10:50:52 crc kubenswrapper[5025]: I1004 10:50:52.635478 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ee98710d-b3c1-4dbf-ab50-ee831c4d622d","Type":"ContainerStarted","Data":"704061e875bd0d83310d6943a8f92ab71e004a1c5c9f89f26f568828f02239fe"} Oct 04 10:50:53 crc kubenswrapper[5025]: I1004 10:50:53.644396 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ee98710d-b3c1-4dbf-ab50-ee831c4d622d","Type":"ContainerStarted","Data":"0cc05f8925e86cef15629a7aabc9f413c8e75fe03105a8fac55dc2c26fe49c8c"} Oct 04 10:50:53 crc kubenswrapper[5025]: I1004 10:50:53.644948 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 04 10:50:53 crc kubenswrapper[5025]: I1004 10:50:53.645044 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ee98710d-b3c1-4dbf-ab50-ee831c4d622d","Type":"ContainerStarted","Data":"7055795d5de346aad25ad76fac3a99145432693389f76defac89c1ebfc6b3f5c"} Oct 04 10:50:53 crc kubenswrapper[5025]: I1004 10:50:53.660147 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.626156921 podStartE2EDuration="2.660128428s" podCreationTimestamp="2025-10-04 10:50:51 +0000 UTC" firstStartedPulling="2025-10-04 10:50:51.967278065 +0000 UTC m=+980.392244935" lastFinishedPulling="2025-10-04 10:50:53.001249562 +0000 UTC m=+981.426216442" observedRunningTime="2025-10-04 10:50:53.659818439 +0000 UTC m=+982.084785319" watchObservedRunningTime="2025-10-04 10:50:53.660128428 +0000 UTC m=+982.085095328" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.281554 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.333153 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.333192 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.346008 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.346079 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.349049 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cszms"] Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.349279 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" podUID="4ac1474e-ef0b-431e-9f1c-45053db21a37" containerName="dnsmasq-dns" containerID="cri-o://e7b213d0a5dcb4c83b558ae5cea062c2c698926a8444a969fa8f6e2b19d88b69" gracePeriod=10 Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.425637 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.425720 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.671882 5025 generic.go:334] "Generic (PLEG): container finished" podID="4ac1474e-ef0b-431e-9f1c-45053db21a37" containerID="e7b213d0a5dcb4c83b558ae5cea062c2c698926a8444a969fa8f6e2b19d88b69" exitCode=0 Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.672794 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" event={"ID":"4ac1474e-ef0b-431e-9f1c-45053db21a37","Type":"ContainerDied","Data":"e7b213d0a5dcb4c83b558ae5cea062c2c698926a8444a969fa8f6e2b19d88b69"} Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.740978 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.741830 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.916156 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.981720 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-dns-svc\") pod \"4ac1474e-ef0b-431e-9f1c-45053db21a37\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.981853 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-config\") pod \"4ac1474e-ef0b-431e-9f1c-45053db21a37\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.981920 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwwtm\" (UniqueName: \"kubernetes.io/projected/4ac1474e-ef0b-431e-9f1c-45053db21a37-kube-api-access-cwwtm\") pod \"4ac1474e-ef0b-431e-9f1c-45053db21a37\" (UID: \"4ac1474e-ef0b-431e-9f1c-45053db21a37\") " Oct 04 10:50:56 crc kubenswrapper[5025]: I1004 10:50:56.988127 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ac1474e-ef0b-431e-9f1c-45053db21a37-kube-api-access-cwwtm" (OuterVolumeSpecName: "kube-api-access-cwwtm") pod "4ac1474e-ef0b-431e-9f1c-45053db21a37" (UID: "4ac1474e-ef0b-431e-9f1c-45053db21a37"). InnerVolumeSpecName "kube-api-access-cwwtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.018282 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4ac1474e-ef0b-431e-9f1c-45053db21a37" (UID: "4ac1474e-ef0b-431e-9f1c-45053db21a37"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.020160 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-config" (OuterVolumeSpecName: "config") pod "4ac1474e-ef0b-431e-9f1c-45053db21a37" (UID: "4ac1474e-ef0b-431e-9f1c-45053db21a37"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.084651 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwwtm\" (UniqueName: \"kubernetes.io/projected/4ac1474e-ef0b-431e-9f1c-45053db21a37-kube-api-access-cwwtm\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.084705 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.084724 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ac1474e-ef0b-431e-9f1c-45053db21a37-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.685291 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" event={"ID":"4ac1474e-ef0b-431e-9f1c-45053db21a37","Type":"ContainerDied","Data":"ac94eb095d8131a712e526bb1a237f5c8dd53fb5163ab1a4a38d3647b6b4b377"} Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.685671 5025 scope.go:117] "RemoveContainer" containerID="e7b213d0a5dcb4c83b558ae5cea062c2c698926a8444a969fa8f6e2b19d88b69" Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.685615 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-cszms" Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.719808 5025 scope.go:117] "RemoveContainer" containerID="945b0c08823efba629a2ae16a6cf74170b83348a2d58213c51c7c829715f4c28" Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.735695 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cszms"] Oct 04 10:50:57 crc kubenswrapper[5025]: I1004 10:50:57.746621 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-cszms"] Oct 04 10:50:58 crc kubenswrapper[5025]: I1004 10:50:58.439987 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ac1474e-ef0b-431e-9f1c-45053db21a37" path="/var/lib/kubelet/pods/4ac1474e-ef0b-431e-9f1c-45053db21a37/volumes" Oct 04 10:50:58 crc kubenswrapper[5025]: I1004 10:50:58.978048 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.131636 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-294dc"] Oct 04 10:50:59 crc kubenswrapper[5025]: E1004 10:50:59.132039 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ac1474e-ef0b-431e-9f1c-45053db21a37" containerName="dnsmasq-dns" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.132054 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ac1474e-ef0b-431e-9f1c-45053db21a37" containerName="dnsmasq-dns" Oct 04 10:50:59 crc kubenswrapper[5025]: E1004 10:50:59.132077 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ac1474e-ef0b-431e-9f1c-45053db21a37" containerName="init" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.132083 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ac1474e-ef0b-431e-9f1c-45053db21a37" containerName="init" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.132251 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ac1474e-ef0b-431e-9f1c-45053db21a37" containerName="dnsmasq-dns" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.138996 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.161314 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-294dc"] Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.241663 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-config\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.241708 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.241734 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cknh\" (UniqueName: \"kubernetes.io/projected/afc4132c-4f96-4898-b18c-4bc79168f8af-kube-api-access-8cknh\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.241755 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.241790 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.343450 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-config\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.343524 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.343555 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cknh\" (UniqueName: \"kubernetes.io/projected/afc4132c-4f96-4898-b18c-4bc79168f8af-kube-api-access-8cknh\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.343586 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.343633 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.344245 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-config\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.344402 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.344603 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.345057 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.363385 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cknh\" (UniqueName: \"kubernetes.io/projected/afc4132c-4f96-4898-b18c-4bc79168f8af-kube-api-access-8cknh\") pod \"dnsmasq-dns-b8fbc5445-294dc\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.457829 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:50:59 crc kubenswrapper[5025]: I1004 10:50:59.888516 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-294dc"] Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.256333 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.264103 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.266248 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.266552 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.268383 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.268433 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-vwscx" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.277724 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.358652 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.358750 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.358792 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-cache\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.358974 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-lock\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.359128 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw9z5\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-kube-api-access-dw9z5\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.460216 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-lock\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.460291 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw9z5\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-kube-api-access-dw9z5\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.460322 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.460368 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.460408 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-cache\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: E1004 10:51:00.460774 5025 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 10:51:00 crc kubenswrapper[5025]: E1004 10:51:00.460793 5025 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 10:51:00 crc kubenswrapper[5025]: E1004 10:51:00.460836 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift podName:369e99d7-ddff-4bfb-9bea-a382aa69eb3e nodeName:}" failed. No retries permitted until 2025-10-04 10:51:00.960819105 +0000 UTC m=+989.385785995 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift") pod "swift-storage-0" (UID: "369e99d7-ddff-4bfb-9bea-a382aa69eb3e") : configmap "swift-ring-files" not found Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.460835 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-cache\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.460920 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.461159 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-lock\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.484315 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw9z5\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-kube-api-access-dw9z5\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.492551 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.712564 5025 generic.go:334] "Generic (PLEG): container finished" podID="afc4132c-4f96-4898-b18c-4bc79168f8af" containerID="ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5" exitCode=0 Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.712622 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" event={"ID":"afc4132c-4f96-4898-b18c-4bc79168f8af","Type":"ContainerDied","Data":"ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5"} Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.712661 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" event={"ID":"afc4132c-4f96-4898-b18c-4bc79168f8af","Type":"ContainerStarted","Data":"0997d5944e57972de5e4cd5138a41314ce4fb1332ec306afe35ba8c44e08f7b6"} Oct 04 10:51:00 crc kubenswrapper[5025]: I1004 10:51:00.967308 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:00 crc kubenswrapper[5025]: E1004 10:51:00.967508 5025 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 10:51:00 crc kubenswrapper[5025]: E1004 10:51:00.967691 5025 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 10:51:00 crc kubenswrapper[5025]: E1004 10:51:00.967752 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift podName:369e99d7-ddff-4bfb-9bea-a382aa69eb3e nodeName:}" failed. No retries permitted until 2025-10-04 10:51:01.96773291 +0000 UTC m=+990.392699790 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift") pod "swift-storage-0" (UID: "369e99d7-ddff-4bfb-9bea-a382aa69eb3e") : configmap "swift-ring-files" not found Oct 04 10:51:01 crc kubenswrapper[5025]: I1004 10:51:01.731471 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" event={"ID":"afc4132c-4f96-4898-b18c-4bc79168f8af","Type":"ContainerStarted","Data":"948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0"} Oct 04 10:51:01 crc kubenswrapper[5025]: I1004 10:51:01.731804 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:51:01 crc kubenswrapper[5025]: I1004 10:51:01.752083 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" podStartSLOduration=2.752055951 podStartE2EDuration="2.752055951s" podCreationTimestamp="2025-10-04 10:50:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:51:01.748091655 +0000 UTC m=+990.173058535" watchObservedRunningTime="2025-10-04 10:51:01.752055951 +0000 UTC m=+990.177022861" Oct 04 10:51:01 crc kubenswrapper[5025]: I1004 10:51:01.990136 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:01 crc kubenswrapper[5025]: E1004 10:51:01.990447 5025 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 10:51:01 crc kubenswrapper[5025]: E1004 10:51:01.990463 5025 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 10:51:01 crc kubenswrapper[5025]: E1004 10:51:01.990506 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift podName:369e99d7-ddff-4bfb-9bea-a382aa69eb3e nodeName:}" failed. No retries permitted until 2025-10-04 10:51:03.990491459 +0000 UTC m=+992.415458349 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift") pod "swift-storage-0" (UID: "369e99d7-ddff-4bfb-9bea-a382aa69eb3e") : configmap "swift-ring-files" not found Oct 04 10:51:02 crc kubenswrapper[5025]: I1004 10:51:02.204493 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-fzr5t"] Oct 04 10:51:02 crc kubenswrapper[5025]: I1004 10:51:02.206083 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fzr5t" Oct 04 10:51:02 crc kubenswrapper[5025]: I1004 10:51:02.215847 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-fzr5t"] Oct 04 10:51:02 crc kubenswrapper[5025]: I1004 10:51:02.294991 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvszd\" (UniqueName: \"kubernetes.io/projected/39ee3358-4ace-4a12-93c8-90f0fbead7c5-kube-api-access-wvszd\") pod \"glance-db-create-fzr5t\" (UID: \"39ee3358-4ace-4a12-93c8-90f0fbead7c5\") " pod="openstack/glance-db-create-fzr5t" Oct 04 10:51:02 crc kubenswrapper[5025]: I1004 10:51:02.397099 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvszd\" (UniqueName: \"kubernetes.io/projected/39ee3358-4ace-4a12-93c8-90f0fbead7c5-kube-api-access-wvszd\") pod \"glance-db-create-fzr5t\" (UID: \"39ee3358-4ace-4a12-93c8-90f0fbead7c5\") " pod="openstack/glance-db-create-fzr5t" Oct 04 10:51:02 crc kubenswrapper[5025]: I1004 10:51:02.438810 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvszd\" (UniqueName: \"kubernetes.io/projected/39ee3358-4ace-4a12-93c8-90f0fbead7c5-kube-api-access-wvszd\") pod \"glance-db-create-fzr5t\" (UID: \"39ee3358-4ace-4a12-93c8-90f0fbead7c5\") " pod="openstack/glance-db-create-fzr5t" Oct 04 10:51:02 crc kubenswrapper[5025]: I1004 10:51:02.529899 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fzr5t" Oct 04 10:51:02 crc kubenswrapper[5025]: I1004 10:51:02.999468 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-fzr5t"] Oct 04 10:51:03 crc kubenswrapper[5025]: W1004 10:51:03.007248 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39ee3358_4ace_4a12_93c8_90f0fbead7c5.slice/crio-a1935b0bb6af952a72c7d7d7fea93fc749c37b5040a5c01a743e2474813b87fa WatchSource:0}: Error finding container a1935b0bb6af952a72c7d7d7fea93fc749c37b5040a5c01a743e2474813b87fa: Status 404 returned error can't find the container with id a1935b0bb6af952a72c7d7d7fea93fc749c37b5040a5c01a743e2474813b87fa Oct 04 10:51:03 crc kubenswrapper[5025]: I1004 10:51:03.748893 5025 generic.go:334] "Generic (PLEG): container finished" podID="39ee3358-4ace-4a12-93c8-90f0fbead7c5" containerID="df5cd369869c48699d0521f44fb93465ad2f421b591b1e128112798d7d497569" exitCode=0 Oct 04 10:51:03 crc kubenswrapper[5025]: I1004 10:51:03.748995 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fzr5t" event={"ID":"39ee3358-4ace-4a12-93c8-90f0fbead7c5","Type":"ContainerDied","Data":"df5cd369869c48699d0521f44fb93465ad2f421b591b1e128112798d7d497569"} Oct 04 10:51:03 crc kubenswrapper[5025]: I1004 10:51:03.749464 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fzr5t" event={"ID":"39ee3358-4ace-4a12-93c8-90f0fbead7c5","Type":"ContainerStarted","Data":"a1935b0bb6af952a72c7d7d7fea93fc749c37b5040a5c01a743e2474813b87fa"} Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.027300 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:04 crc kubenswrapper[5025]: E1004 10:51:04.027606 5025 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 10:51:04 crc kubenswrapper[5025]: E1004 10:51:04.027644 5025 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 10:51:04 crc kubenswrapper[5025]: E1004 10:51:04.027740 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift podName:369e99d7-ddff-4bfb-9bea-a382aa69eb3e nodeName:}" failed. No retries permitted until 2025-10-04 10:51:08.027709646 +0000 UTC m=+996.452676556 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift") pod "swift-storage-0" (UID: "369e99d7-ddff-4bfb-9bea-a382aa69eb3e") : configmap "swift-ring-files" not found Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.166081 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-wmvlp"] Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.167604 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.170145 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.170429 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.170588 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.187898 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wmvlp"] Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.230373 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/49009ba5-e7dd-408f-ad33-a9fa20d156a6-etc-swift\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.230452 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-dispersionconf\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.230633 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-swiftconf\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.230769 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-ring-data-devices\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.230808 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-scripts\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.230838 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-combined-ca-bundle\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.230919 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q2p8\" (UniqueName: \"kubernetes.io/projected/49009ba5-e7dd-408f-ad33-a9fa20d156a6-kube-api-access-4q2p8\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.332452 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-ring-data-devices\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.332531 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-scripts\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.332632 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-combined-ca-bundle\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.332746 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q2p8\" (UniqueName: \"kubernetes.io/projected/49009ba5-e7dd-408f-ad33-a9fa20d156a6-kube-api-access-4q2p8\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.332795 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/49009ba5-e7dd-408f-ad33-a9fa20d156a6-etc-swift\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.332847 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-dispersionconf\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.332948 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-swiftconf\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.333728 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/49009ba5-e7dd-408f-ad33-a9fa20d156a6-etc-swift\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.334304 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-ring-data-devices\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.334604 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-scripts\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.339948 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-dispersionconf\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.340720 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-combined-ca-bundle\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.341937 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-swiftconf\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.364992 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q2p8\" (UniqueName: \"kubernetes.io/projected/49009ba5-e7dd-408f-ad33-a9fa20d156a6-kube-api-access-4q2p8\") pod \"swift-ring-rebalance-wmvlp\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.487528 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:04 crc kubenswrapper[5025]: I1004 10:51:04.958544 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-wmvlp"] Oct 04 10:51:04 crc kubenswrapper[5025]: W1004 10:51:04.963149 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49009ba5_e7dd_408f_ad33_a9fa20d156a6.slice/crio-1260590f2a3f85fb266912e0bf32b87fdef4fb65e13c46944e8d4052f5798e89 WatchSource:0}: Error finding container 1260590f2a3f85fb266912e0bf32b87fdef4fb65e13c46944e8d4052f5798e89: Status 404 returned error can't find the container with id 1260590f2a3f85fb266912e0bf32b87fdef4fb65e13c46944e8d4052f5798e89 Oct 04 10:51:05 crc kubenswrapper[5025]: I1004 10:51:05.120042 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fzr5t" Oct 04 10:51:05 crc kubenswrapper[5025]: I1004 10:51:05.249981 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvszd\" (UniqueName: \"kubernetes.io/projected/39ee3358-4ace-4a12-93c8-90f0fbead7c5-kube-api-access-wvszd\") pod \"39ee3358-4ace-4a12-93c8-90f0fbead7c5\" (UID: \"39ee3358-4ace-4a12-93c8-90f0fbead7c5\") " Oct 04 10:51:05 crc kubenswrapper[5025]: I1004 10:51:05.254729 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39ee3358-4ace-4a12-93c8-90f0fbead7c5-kube-api-access-wvszd" (OuterVolumeSpecName: "kube-api-access-wvszd") pod "39ee3358-4ace-4a12-93c8-90f0fbead7c5" (UID: "39ee3358-4ace-4a12-93c8-90f0fbead7c5"). InnerVolumeSpecName "kube-api-access-wvszd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:05 crc kubenswrapper[5025]: I1004 10:51:05.352268 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvszd\" (UniqueName: \"kubernetes.io/projected/39ee3358-4ace-4a12-93c8-90f0fbead7c5-kube-api-access-wvszd\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:05 crc kubenswrapper[5025]: I1004 10:51:05.768412 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wmvlp" event={"ID":"49009ba5-e7dd-408f-ad33-a9fa20d156a6","Type":"ContainerStarted","Data":"1260590f2a3f85fb266912e0bf32b87fdef4fb65e13c46944e8d4052f5798e89"} Oct 04 10:51:05 crc kubenswrapper[5025]: I1004 10:51:05.770874 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-fzr5t" event={"ID":"39ee3358-4ace-4a12-93c8-90f0fbead7c5","Type":"ContainerDied","Data":"a1935b0bb6af952a72c7d7d7fea93fc749c37b5040a5c01a743e2474813b87fa"} Oct 04 10:51:05 crc kubenswrapper[5025]: I1004 10:51:05.770909 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1935b0bb6af952a72c7d7d7fea93fc749c37b5040a5c01a743e2474813b87fa" Oct 04 10:51:05 crc kubenswrapper[5025]: I1004 10:51:05.770966 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-fzr5t" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.533539 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.551823 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-jhjgk"] Oct 04 10:51:06 crc kubenswrapper[5025]: E1004 10:51:06.552414 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ee3358-4ace-4a12-93c8-90f0fbead7c5" containerName="mariadb-database-create" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.552438 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ee3358-4ace-4a12-93c8-90f0fbead7c5" containerName="mariadb-database-create" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.552840 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ee3358-4ace-4a12-93c8-90f0fbead7c5" containerName="mariadb-database-create" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.554204 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jhjgk" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.589468 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jhjgk"] Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.675815 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf4bv\" (UniqueName: \"kubernetes.io/projected/8c58f978-1c60-4b77-b6cc-519d0b28f447-kube-api-access-lf4bv\") pod \"keystone-db-create-jhjgk\" (UID: \"8c58f978-1c60-4b77-b6cc-519d0b28f447\") " pod="openstack/keystone-db-create-jhjgk" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.778066 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf4bv\" (UniqueName: \"kubernetes.io/projected/8c58f978-1c60-4b77-b6cc-519d0b28f447-kube-api-access-lf4bv\") pod \"keystone-db-create-jhjgk\" (UID: \"8c58f978-1c60-4b77-b6cc-519d0b28f447\") " pod="openstack/keystone-db-create-jhjgk" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.796441 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf4bv\" (UniqueName: \"kubernetes.io/projected/8c58f978-1c60-4b77-b6cc-519d0b28f447-kube-api-access-lf4bv\") pod \"keystone-db-create-jhjgk\" (UID: \"8c58f978-1c60-4b77-b6cc-519d0b28f447\") " pod="openstack/keystone-db-create-jhjgk" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.832356 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-k4fff"] Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.833598 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-k4fff" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.839558 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-k4fff"] Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.880547 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jhjgk" Oct 04 10:51:06 crc kubenswrapper[5025]: I1004 10:51:06.982580 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2rmd\" (UniqueName: \"kubernetes.io/projected/ac1e5058-70ea-478e-b265-d737830e6daa-kube-api-access-k2rmd\") pod \"placement-db-create-k4fff\" (UID: \"ac1e5058-70ea-478e-b265-d737830e6daa\") " pod="openstack/placement-db-create-k4fff" Oct 04 10:51:07 crc kubenswrapper[5025]: I1004 10:51:07.083894 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2rmd\" (UniqueName: \"kubernetes.io/projected/ac1e5058-70ea-478e-b265-d737830e6daa-kube-api-access-k2rmd\") pod \"placement-db-create-k4fff\" (UID: \"ac1e5058-70ea-478e-b265-d737830e6daa\") " pod="openstack/placement-db-create-k4fff" Oct 04 10:51:07 crc kubenswrapper[5025]: I1004 10:51:07.101998 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2rmd\" (UniqueName: \"kubernetes.io/projected/ac1e5058-70ea-478e-b265-d737830e6daa-kube-api-access-k2rmd\") pod \"placement-db-create-k4fff\" (UID: \"ac1e5058-70ea-478e-b265-d737830e6daa\") " pod="openstack/placement-db-create-k4fff" Oct 04 10:51:07 crc kubenswrapper[5025]: I1004 10:51:07.156886 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-k4fff" Oct 04 10:51:07 crc kubenswrapper[5025]: I1004 10:51:07.797427 5025 generic.go:334] "Generic (PLEG): container finished" podID="a36a09c8-4254-4f14-bd39-7156cb462adb" containerID="66815974cd0d2555f9c56c6b7046d5ea1aa08f08cf16377d7c60422b077b160a" exitCode=0 Oct 04 10:51:07 crc kubenswrapper[5025]: I1004 10:51:07.797503 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a36a09c8-4254-4f14-bd39-7156cb462adb","Type":"ContainerDied","Data":"66815974cd0d2555f9c56c6b7046d5ea1aa08f08cf16377d7c60422b077b160a"} Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.105236 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:08 crc kubenswrapper[5025]: E1004 10:51:08.105771 5025 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 04 10:51:08 crc kubenswrapper[5025]: E1004 10:51:08.105788 5025 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 04 10:51:08 crc kubenswrapper[5025]: E1004 10:51:08.105842 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift podName:369e99d7-ddff-4bfb-9bea-a382aa69eb3e nodeName:}" failed. No retries permitted until 2025-10-04 10:51:16.105821867 +0000 UTC m=+1004.530788757 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift") pod "swift-storage-0" (UID: "369e99d7-ddff-4bfb-9bea-a382aa69eb3e") : configmap "swift-ring-files" not found Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.345206 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-k4fff"] Oct 04 10:51:08 crc kubenswrapper[5025]: W1004 10:51:08.353674 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac1e5058_70ea_478e_b265_d737830e6daa.slice/crio-763dadcace235318be4b27a9e273208ff70701282ddc3304c605b40cc82d782a WatchSource:0}: Error finding container 763dadcace235318be4b27a9e273208ff70701282ddc3304c605b40cc82d782a: Status 404 returned error can't find the container with id 763dadcace235318be4b27a9e273208ff70701282ddc3304c605b40cc82d782a Oct 04 10:51:08 crc kubenswrapper[5025]: W1004 10:51:08.447852 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c58f978_1c60_4b77_b6cc_519d0b28f447.slice/crio-12eb2e1ec437656bf77e6045fe2195b42d99101a0813c432cbb34112b4e5d994 WatchSource:0}: Error finding container 12eb2e1ec437656bf77e6045fe2195b42d99101a0813c432cbb34112b4e5d994: Status 404 returned error can't find the container with id 12eb2e1ec437656bf77e6045fe2195b42d99101a0813c432cbb34112b4e5d994 Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.450453 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-jhjgk"] Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.807732 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wmvlp" event={"ID":"49009ba5-e7dd-408f-ad33-a9fa20d156a6","Type":"ContainerStarted","Data":"3028499b1dae5a2a3a95e8f79b5d5a4f3ca9b58e293869ed95b819324294a4a6"} Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.809474 5025 generic.go:334] "Generic (PLEG): container finished" podID="ac1e5058-70ea-478e-b265-d737830e6daa" containerID="f4b369e9a8aabe353e58f5866e604c4783c6390252fa27d24258c2e9b3b60e43" exitCode=0 Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.809565 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-k4fff" event={"ID":"ac1e5058-70ea-478e-b265-d737830e6daa","Type":"ContainerDied","Data":"f4b369e9a8aabe353e58f5866e604c4783c6390252fa27d24258c2e9b3b60e43"} Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.809705 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-k4fff" event={"ID":"ac1e5058-70ea-478e-b265-d737830e6daa","Type":"ContainerStarted","Data":"763dadcace235318be4b27a9e273208ff70701282ddc3304c605b40cc82d782a"} Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.812110 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a36a09c8-4254-4f14-bd39-7156cb462adb","Type":"ContainerStarted","Data":"2ab8fc08c3ee477353e833ec543df74b381fd508f89e05deb084314d2a145f33"} Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.813148 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.815348 5025 generic.go:334] "Generic (PLEG): container finished" podID="8c58f978-1c60-4b77-b6cc-519d0b28f447" containerID="19bbbec6b4c169ccdae4e47b4e0dcbfaf353e25657dce3380401556152a324fc" exitCode=0 Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.815505 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jhjgk" event={"ID":"8c58f978-1c60-4b77-b6cc-519d0b28f447","Type":"ContainerDied","Data":"19bbbec6b4c169ccdae4e47b4e0dcbfaf353e25657dce3380401556152a324fc"} Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.815590 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jhjgk" event={"ID":"8c58f978-1c60-4b77-b6cc-519d0b28f447","Type":"ContainerStarted","Data":"12eb2e1ec437656bf77e6045fe2195b42d99101a0813c432cbb34112b4e5d994"} Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.817542 5025 generic.go:334] "Generic (PLEG): container finished" podID="9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" containerID="2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea" exitCode=0 Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.817592 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a","Type":"ContainerDied","Data":"2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea"} Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.837543 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-wmvlp" podStartSLOduration=1.8976411949999998 podStartE2EDuration="4.837514561s" podCreationTimestamp="2025-10-04 10:51:04 +0000 UTC" firstStartedPulling="2025-10-04 10:51:04.965549164 +0000 UTC m=+993.390516044" lastFinishedPulling="2025-10-04 10:51:07.90542253 +0000 UTC m=+996.330389410" observedRunningTime="2025-10-04 10:51:08.828202888 +0000 UTC m=+997.253169838" watchObservedRunningTime="2025-10-04 10:51:08.837514561 +0000 UTC m=+997.262481481" Oct 04 10:51:08 crc kubenswrapper[5025]: I1004 10:51:08.982879 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.909763904 podStartE2EDuration="56.982853598s" podCreationTimestamp="2025-10-04 10:50:12 +0000 UTC" firstStartedPulling="2025-10-04 10:50:14.027359294 +0000 UTC m=+942.452326174" lastFinishedPulling="2025-10-04 10:50:34.100448958 +0000 UTC m=+962.525415868" observedRunningTime="2025-10-04 10:51:08.978526552 +0000 UTC m=+997.403493432" watchObservedRunningTime="2025-10-04 10:51:08.982853598 +0000 UTC m=+997.407820478" Oct 04 10:51:09 crc kubenswrapper[5025]: I1004 10:51:09.459250 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:51:09 crc kubenswrapper[5025]: I1004 10:51:09.519331 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnsqw"] Oct 04 10:51:09 crc kubenswrapper[5025]: I1004 10:51:09.519614 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-mnsqw" podUID="49e49e17-d84c-4614-8e56-02049a145e26" containerName="dnsmasq-dns" containerID="cri-o://054a8e988af91704459b80f6746d1324e8ca08b74b076cb6e057906b86bf0906" gracePeriod=10 Oct 04 10:51:09 crc kubenswrapper[5025]: I1004 10:51:09.828150 5025 generic.go:334] "Generic (PLEG): container finished" podID="49e49e17-d84c-4614-8e56-02049a145e26" containerID="054a8e988af91704459b80f6746d1324e8ca08b74b076cb6e057906b86bf0906" exitCode=0 Oct 04 10:51:09 crc kubenswrapper[5025]: I1004 10:51:09.828416 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-mnsqw" event={"ID":"49e49e17-d84c-4614-8e56-02049a145e26","Type":"ContainerDied","Data":"054a8e988af91704459b80f6746d1324e8ca08b74b076cb6e057906b86bf0906"} Oct 04 10:51:09 crc kubenswrapper[5025]: I1004 10:51:09.831051 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a","Type":"ContainerStarted","Data":"9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86"} Oct 04 10:51:09 crc kubenswrapper[5025]: I1004 10:51:09.832247 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.012421 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.034969 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=42.301066667 podStartE2EDuration="58.034953914s" podCreationTimestamp="2025-10-04 10:50:12 +0000 UTC" firstStartedPulling="2025-10-04 10:50:18.812276084 +0000 UTC m=+947.237242964" lastFinishedPulling="2025-10-04 10:50:34.546163311 +0000 UTC m=+962.971130211" observedRunningTime="2025-10-04 10:51:09.867295355 +0000 UTC m=+998.292262235" watchObservedRunningTime="2025-10-04 10:51:10.034953914 +0000 UTC m=+998.459920794" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.081800 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-sb\") pod \"49e49e17-d84c-4614-8e56-02049a145e26\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.081936 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cn6t\" (UniqueName: \"kubernetes.io/projected/49e49e17-d84c-4614-8e56-02049a145e26-kube-api-access-5cn6t\") pod \"49e49e17-d84c-4614-8e56-02049a145e26\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.082073 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-config\") pod \"49e49e17-d84c-4614-8e56-02049a145e26\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.082110 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-nb\") pod \"49e49e17-d84c-4614-8e56-02049a145e26\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.082153 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-dns-svc\") pod \"49e49e17-d84c-4614-8e56-02049a145e26\" (UID: \"49e49e17-d84c-4614-8e56-02049a145e26\") " Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.114513 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49e49e17-d84c-4614-8e56-02049a145e26-kube-api-access-5cn6t" (OuterVolumeSpecName: "kube-api-access-5cn6t") pod "49e49e17-d84c-4614-8e56-02049a145e26" (UID: "49e49e17-d84c-4614-8e56-02049a145e26"). InnerVolumeSpecName "kube-api-access-5cn6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.147826 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "49e49e17-d84c-4614-8e56-02049a145e26" (UID: "49e49e17-d84c-4614-8e56-02049a145e26"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.171716 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-config" (OuterVolumeSpecName: "config") pod "49e49e17-d84c-4614-8e56-02049a145e26" (UID: "49e49e17-d84c-4614-8e56-02049a145e26"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.177595 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "49e49e17-d84c-4614-8e56-02049a145e26" (UID: "49e49e17-d84c-4614-8e56-02049a145e26"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.177615 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "49e49e17-d84c-4614-8e56-02049a145e26" (UID: "49e49e17-d84c-4614-8e56-02049a145e26"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.184288 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.184327 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.184340 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.184353 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/49e49e17-d84c-4614-8e56-02049a145e26-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.184367 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cn6t\" (UniqueName: \"kubernetes.io/projected/49e49e17-d84c-4614-8e56-02049a145e26-kube-api-access-5cn6t\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.249745 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jhjgk" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.271481 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-k4fff" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.391717 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2rmd\" (UniqueName: \"kubernetes.io/projected/ac1e5058-70ea-478e-b265-d737830e6daa-kube-api-access-k2rmd\") pod \"ac1e5058-70ea-478e-b265-d737830e6daa\" (UID: \"ac1e5058-70ea-478e-b265-d737830e6daa\") " Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.391895 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf4bv\" (UniqueName: \"kubernetes.io/projected/8c58f978-1c60-4b77-b6cc-519d0b28f447-kube-api-access-lf4bv\") pod \"8c58f978-1c60-4b77-b6cc-519d0b28f447\" (UID: \"8c58f978-1c60-4b77-b6cc-519d0b28f447\") " Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.395550 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c58f978-1c60-4b77-b6cc-519d0b28f447-kube-api-access-lf4bv" (OuterVolumeSpecName: "kube-api-access-lf4bv") pod "8c58f978-1c60-4b77-b6cc-519d0b28f447" (UID: "8c58f978-1c60-4b77-b6cc-519d0b28f447"). InnerVolumeSpecName "kube-api-access-lf4bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.395656 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac1e5058-70ea-478e-b265-d737830e6daa-kube-api-access-k2rmd" (OuterVolumeSpecName: "kube-api-access-k2rmd") pod "ac1e5058-70ea-478e-b265-d737830e6daa" (UID: "ac1e5058-70ea-478e-b265-d737830e6daa"). InnerVolumeSpecName "kube-api-access-k2rmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.493820 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2rmd\" (UniqueName: \"kubernetes.io/projected/ac1e5058-70ea-478e-b265-d737830e6daa-kube-api-access-k2rmd\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.493851 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf4bv\" (UniqueName: \"kubernetes.io/projected/8c58f978-1c60-4b77-b6cc-519d0b28f447-kube-api-access-lf4bv\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.840392 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-k4fff" event={"ID":"ac1e5058-70ea-478e-b265-d737830e6daa","Type":"ContainerDied","Data":"763dadcace235318be4b27a9e273208ff70701282ddc3304c605b40cc82d782a"} Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.840441 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="763dadcace235318be4b27a9e273208ff70701282ddc3304c605b40cc82d782a" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.840522 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-k4fff" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.844001 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-mnsqw" event={"ID":"49e49e17-d84c-4614-8e56-02049a145e26","Type":"ContainerDied","Data":"dd887aaa060fbb1edbf182b5a38f6cd44c2bb26964691d6d9006a2145d862d69"} Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.844071 5025 scope.go:117] "RemoveContainer" containerID="054a8e988af91704459b80f6746d1324e8ca08b74b076cb6e057906b86bf0906" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.844203 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-mnsqw" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.849194 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-jhjgk" event={"ID":"8c58f978-1c60-4b77-b6cc-519d0b28f447","Type":"ContainerDied","Data":"12eb2e1ec437656bf77e6045fe2195b42d99101a0813c432cbb34112b4e5d994"} Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.849252 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12eb2e1ec437656bf77e6045fe2195b42d99101a0813c432cbb34112b4e5d994" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.849482 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-jhjgk" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.877700 5025 scope.go:117] "RemoveContainer" containerID="f4bd2a72b24d93cfa6908f1b9b80e79d7b0248f2e18f92c45a657e46a046d141" Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.883622 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnsqw"] Oct 04 10:51:10 crc kubenswrapper[5025]: I1004 10:51:10.894137 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-mnsqw"] Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.232859 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-e3d9-account-create-hrch9"] Oct 04 10:51:12 crc kubenswrapper[5025]: E1004 10:51:12.233291 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c58f978-1c60-4b77-b6cc-519d0b28f447" containerName="mariadb-database-create" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.233309 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c58f978-1c60-4b77-b6cc-519d0b28f447" containerName="mariadb-database-create" Oct 04 10:51:12 crc kubenswrapper[5025]: E1004 10:51:12.233329 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e49e17-d84c-4614-8e56-02049a145e26" containerName="init" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.233337 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e49e17-d84c-4614-8e56-02049a145e26" containerName="init" Oct 04 10:51:12 crc kubenswrapper[5025]: E1004 10:51:12.233357 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49e49e17-d84c-4614-8e56-02049a145e26" containerName="dnsmasq-dns" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.233366 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="49e49e17-d84c-4614-8e56-02049a145e26" containerName="dnsmasq-dns" Oct 04 10:51:12 crc kubenswrapper[5025]: E1004 10:51:12.233379 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1e5058-70ea-478e-b265-d737830e6daa" containerName="mariadb-database-create" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.233387 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1e5058-70ea-478e-b265-d737830e6daa" containerName="mariadb-database-create" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.233573 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c58f978-1c60-4b77-b6cc-519d0b28f447" containerName="mariadb-database-create" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.233592 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="49e49e17-d84c-4614-8e56-02049a145e26" containerName="dnsmasq-dns" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.233606 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac1e5058-70ea-478e-b265-d737830e6daa" containerName="mariadb-database-create" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.234188 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e3d9-account-create-hrch9" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.240137 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.249032 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e3d9-account-create-hrch9"] Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.326305 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r2md\" (UniqueName: \"kubernetes.io/projected/06c89e30-b6f2-4416-a976-e0661bd1fa07-kube-api-access-4r2md\") pod \"glance-e3d9-account-create-hrch9\" (UID: \"06c89e30-b6f2-4416-a976-e0661bd1fa07\") " pod="openstack/glance-e3d9-account-create-hrch9" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.422752 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49e49e17-d84c-4614-8e56-02049a145e26" path="/var/lib/kubelet/pods/49e49e17-d84c-4614-8e56-02049a145e26/volumes" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.428308 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r2md\" (UniqueName: \"kubernetes.io/projected/06c89e30-b6f2-4416-a976-e0661bd1fa07-kube-api-access-4r2md\") pod \"glance-e3d9-account-create-hrch9\" (UID: \"06c89e30-b6f2-4416-a976-e0661bd1fa07\") " pod="openstack/glance-e3d9-account-create-hrch9" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.451697 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r2md\" (UniqueName: \"kubernetes.io/projected/06c89e30-b6f2-4416-a976-e0661bd1fa07-kube-api-access-4r2md\") pod \"glance-e3d9-account-create-hrch9\" (UID: \"06c89e30-b6f2-4416-a976-e0661bd1fa07\") " pod="openstack/glance-e3d9-account-create-hrch9" Oct 04 10:51:12 crc kubenswrapper[5025]: I1004 10:51:12.599434 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e3d9-account-create-hrch9" Oct 04 10:51:13 crc kubenswrapper[5025]: I1004 10:51:13.079001 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e3d9-account-create-hrch9"] Oct 04 10:51:13 crc kubenswrapper[5025]: W1004 10:51:13.081339 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06c89e30_b6f2_4416_a976_e0661bd1fa07.slice/crio-a0f42f7d0f2d8f002ac9807875881ab09c6319ccd9c2ffd9b14cc9faa87cec78 WatchSource:0}: Error finding container a0f42f7d0f2d8f002ac9807875881ab09c6319ccd9c2ffd9b14cc9faa87cec78: Status 404 returned error can't find the container with id a0f42f7d0f2d8f002ac9807875881ab09c6319ccd9c2ffd9b14cc9faa87cec78 Oct 04 10:51:13 crc kubenswrapper[5025]: I1004 10:51:13.884628 5025 generic.go:334] "Generic (PLEG): container finished" podID="06c89e30-b6f2-4416-a976-e0661bd1fa07" containerID="df0305ddf1cd74da92eed6ad7d1953ed912edde49e3d7488ae3ed3e9baa705c7" exitCode=0 Oct 04 10:51:13 crc kubenswrapper[5025]: I1004 10:51:13.884681 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e3d9-account-create-hrch9" event={"ID":"06c89e30-b6f2-4416-a976-e0661bd1fa07","Type":"ContainerDied","Data":"df0305ddf1cd74da92eed6ad7d1953ed912edde49e3d7488ae3ed3e9baa705c7"} Oct 04 10:51:13 crc kubenswrapper[5025]: I1004 10:51:13.884713 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e3d9-account-create-hrch9" event={"ID":"06c89e30-b6f2-4416-a976-e0661bd1fa07","Type":"ContainerStarted","Data":"a0f42f7d0f2d8f002ac9807875881ab09c6319ccd9c2ffd9b14cc9faa87cec78"} Oct 04 10:51:14 crc kubenswrapper[5025]: I1004 10:51:14.713904 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:51:14 crc kubenswrapper[5025]: I1004 10:51:14.714574 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:51:15 crc kubenswrapper[5025]: I1004 10:51:15.273898 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e3d9-account-create-hrch9" Oct 04 10:51:15 crc kubenswrapper[5025]: I1004 10:51:15.371250 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r2md\" (UniqueName: \"kubernetes.io/projected/06c89e30-b6f2-4416-a976-e0661bd1fa07-kube-api-access-4r2md\") pod \"06c89e30-b6f2-4416-a976-e0661bd1fa07\" (UID: \"06c89e30-b6f2-4416-a976-e0661bd1fa07\") " Oct 04 10:51:15 crc kubenswrapper[5025]: I1004 10:51:15.379211 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06c89e30-b6f2-4416-a976-e0661bd1fa07-kube-api-access-4r2md" (OuterVolumeSpecName: "kube-api-access-4r2md") pod "06c89e30-b6f2-4416-a976-e0661bd1fa07" (UID: "06c89e30-b6f2-4416-a976-e0661bd1fa07"). InnerVolumeSpecName "kube-api-access-4r2md". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:15 crc kubenswrapper[5025]: I1004 10:51:15.473199 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r2md\" (UniqueName: \"kubernetes.io/projected/06c89e30-b6f2-4416-a976-e0661bd1fa07-kube-api-access-4r2md\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:15 crc kubenswrapper[5025]: I1004 10:51:15.899817 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e3d9-account-create-hrch9" Oct 04 10:51:15 crc kubenswrapper[5025]: I1004 10:51:15.899827 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e3d9-account-create-hrch9" event={"ID":"06c89e30-b6f2-4416-a976-e0661bd1fa07","Type":"ContainerDied","Data":"a0f42f7d0f2d8f002ac9807875881ab09c6319ccd9c2ffd9b14cc9faa87cec78"} Oct 04 10:51:15 crc kubenswrapper[5025]: I1004 10:51:15.899903 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0f42f7d0f2d8f002ac9807875881ab09c6319ccd9c2ffd9b14cc9faa87cec78" Oct 04 10:51:15 crc kubenswrapper[5025]: I1004 10:51:15.901988 5025 generic.go:334] "Generic (PLEG): container finished" podID="49009ba5-e7dd-408f-ad33-a9fa20d156a6" containerID="3028499b1dae5a2a3a95e8f79b5d5a4f3ca9b58e293869ed95b819324294a4a6" exitCode=0 Oct 04 10:51:15 crc kubenswrapper[5025]: I1004 10:51:15.902047 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wmvlp" event={"ID":"49009ba5-e7dd-408f-ad33-a9fa20d156a6","Type":"ContainerDied","Data":"3028499b1dae5a2a3a95e8f79b5d5a4f3ca9b58e293869ed95b819324294a4a6"} Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.185648 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.191146 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/369e99d7-ddff-4bfb-9bea-a382aa69eb3e-etc-swift\") pod \"swift-storage-0\" (UID: \"369e99d7-ddff-4bfb-9bea-a382aa69eb3e\") " pod="openstack/swift-storage-0" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.193347 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.705953 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-ccc4-account-create-pl9j2"] Oct 04 10:51:16 crc kubenswrapper[5025]: E1004 10:51:16.706791 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c89e30-b6f2-4416-a976-e0661bd1fa07" containerName="mariadb-account-create" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.706809 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c89e30-b6f2-4416-a976-e0661bd1fa07" containerName="mariadb-account-create" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.707025 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="06c89e30-b6f2-4416-a976-e0661bd1fa07" containerName="mariadb-account-create" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.707776 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ccc4-account-create-pl9j2" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.709715 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.715846 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ccc4-account-create-pl9j2"] Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.791243 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.819092 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkkmv\" (UniqueName: \"kubernetes.io/projected/dad9687f-030b-4379-ba71-965b512fadaa-kube-api-access-rkkmv\") pod \"keystone-ccc4-account-create-pl9j2\" (UID: \"dad9687f-030b-4379-ba71-965b512fadaa\") " pod="openstack/keystone-ccc4-account-create-pl9j2" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.914668 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"1cdccd93357b9c44f58268b7332a057d1920c27f6e9a922497e2f512dea228aa"} Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.921248 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkkmv\" (UniqueName: \"kubernetes.io/projected/dad9687f-030b-4379-ba71-965b512fadaa-kube-api-access-rkkmv\") pod \"keystone-ccc4-account-create-pl9j2\" (UID: \"dad9687f-030b-4379-ba71-965b512fadaa\") " pod="openstack/keystone-ccc4-account-create-pl9j2" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.945535 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkkmv\" (UniqueName: \"kubernetes.io/projected/dad9687f-030b-4379-ba71-965b512fadaa-kube-api-access-rkkmv\") pod \"keystone-ccc4-account-create-pl9j2\" (UID: \"dad9687f-030b-4379-ba71-965b512fadaa\") " pod="openstack/keystone-ccc4-account-create-pl9j2" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.978626 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-9b7d-account-create-q2442"] Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.979668 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9b7d-account-create-q2442" Oct 04 10:51:16 crc kubenswrapper[5025]: I1004 10:51:16.983439 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.005595 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9b7d-account-create-q2442"] Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.029093 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ccc4-account-create-pl9j2" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.123435 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m7jc\" (UniqueName: \"kubernetes.io/projected/083601ee-774b-4e6d-a8e0-debebeb9011a-kube-api-access-8m7jc\") pod \"placement-9b7d-account-create-q2442\" (UID: \"083601ee-774b-4e6d-a8e0-debebeb9011a\") " pod="openstack/placement-9b7d-account-create-q2442" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.197633 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.230384 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m7jc\" (UniqueName: \"kubernetes.io/projected/083601ee-774b-4e6d-a8e0-debebeb9011a-kube-api-access-8m7jc\") pod \"placement-9b7d-account-create-q2442\" (UID: \"083601ee-774b-4e6d-a8e0-debebeb9011a\") " pod="openstack/placement-9b7d-account-create-q2442" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.249174 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m7jc\" (UniqueName: \"kubernetes.io/projected/083601ee-774b-4e6d-a8e0-debebeb9011a-kube-api-access-8m7jc\") pod \"placement-9b7d-account-create-q2442\" (UID: \"083601ee-774b-4e6d-a8e0-debebeb9011a\") " pod="openstack/placement-9b7d-account-create-q2442" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.315937 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9b7d-account-create-q2442" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.332992 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-scripts\") pod \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.333068 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-combined-ca-bundle\") pod \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.333200 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/49009ba5-e7dd-408f-ad33-a9fa20d156a6-etc-swift\") pod \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.333230 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-swiftconf\") pod \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.333262 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4q2p8\" (UniqueName: \"kubernetes.io/projected/49009ba5-e7dd-408f-ad33-a9fa20d156a6-kube-api-access-4q2p8\") pod \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.334684 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-ring-data-devices\") pod \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.334747 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-dispersionconf\") pod \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\" (UID: \"49009ba5-e7dd-408f-ad33-a9fa20d156a6\") " Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.338535 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49009ba5-e7dd-408f-ad33-a9fa20d156a6-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "49009ba5-e7dd-408f-ad33-a9fa20d156a6" (UID: "49009ba5-e7dd-408f-ad33-a9fa20d156a6"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.340562 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "49009ba5-e7dd-408f-ad33-a9fa20d156a6" (UID: "49009ba5-e7dd-408f-ad33-a9fa20d156a6"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.359890 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49009ba5-e7dd-408f-ad33-a9fa20d156a6-kube-api-access-4q2p8" (OuterVolumeSpecName: "kube-api-access-4q2p8") pod "49009ba5-e7dd-408f-ad33-a9fa20d156a6" (UID: "49009ba5-e7dd-408f-ad33-a9fa20d156a6"). InnerVolumeSpecName "kube-api-access-4q2p8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.362594 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "49009ba5-e7dd-408f-ad33-a9fa20d156a6" (UID: "49009ba5-e7dd-408f-ad33-a9fa20d156a6"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.392683 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-frhrp"] Oct 04 10:51:17 crc kubenswrapper[5025]: E1004 10:51:17.393194 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49009ba5-e7dd-408f-ad33-a9fa20d156a6" containerName="swift-ring-rebalance" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.393221 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="49009ba5-e7dd-408f-ad33-a9fa20d156a6" containerName="swift-ring-rebalance" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.393415 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="49009ba5-e7dd-408f-ad33-a9fa20d156a6" containerName="swift-ring-rebalance" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.394073 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.394171 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "49009ba5-e7dd-408f-ad33-a9fa20d156a6" (UID: "49009ba5-e7dd-408f-ad33-a9fa20d156a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.396318 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4867c" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.396527 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.399682 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-frhrp"] Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.402844 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-scripts" (OuterVolumeSpecName: "scripts") pod "49009ba5-e7dd-408f-ad33-a9fa20d156a6" (UID: "49009ba5-e7dd-408f-ad33-a9fa20d156a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.404892 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "49009ba5-e7dd-408f-ad33-a9fa20d156a6" (UID: "49009ba5-e7dd-408f-ad33-a9fa20d156a6"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.437137 5025 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/49009ba5-e7dd-408f-ad33-a9fa20d156a6-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.437178 5025 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.437190 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4q2p8\" (UniqueName: \"kubernetes.io/projected/49009ba5-e7dd-408f-ad33-a9fa20d156a6-kube-api-access-4q2p8\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.437202 5025 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.437222 5025 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.437233 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/49009ba5-e7dd-408f-ad33-a9fa20d156a6-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.437244 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49009ba5-e7dd-408f-ad33-a9fa20d156a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.539323 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl6rl\" (UniqueName: \"kubernetes.io/projected/4d3aea52-35ad-49e4-93fb-465cd111f845-kube-api-access-cl6rl\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.539892 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-config-data\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.540475 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ccc4-account-create-pl9j2"] Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.540413 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-combined-ca-bundle\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.541193 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-db-sync-config-data\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.643353 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-combined-ca-bundle\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.643394 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-db-sync-config-data\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.643468 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl6rl\" (UniqueName: \"kubernetes.io/projected/4d3aea52-35ad-49e4-93fb-465cd111f845-kube-api-access-cl6rl\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.643504 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-config-data\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.649437 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-config-data\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.651612 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-combined-ca-bundle\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.652738 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-db-sync-config-data\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.661155 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl6rl\" (UniqueName: \"kubernetes.io/projected/4d3aea52-35ad-49e4-93fb-465cd111f845-kube-api-access-cl6rl\") pod \"glance-db-sync-frhrp\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.781030 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.808327 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-9b7d-account-create-q2442"] Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.809233 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jcfxn" podUID="fe21ee2a-4ed1-47aa-90f3-42629a279fd6" containerName="ovn-controller" probeResult="failure" output=< Oct 04 10:51:17 crc kubenswrapper[5025]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 04 10:51:17 crc kubenswrapper[5025]: > Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.814284 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.830582 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-c7fwx" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.924749 5025 generic.go:334] "Generic (PLEG): container finished" podID="dad9687f-030b-4379-ba71-965b512fadaa" containerID="d66c64aa24974228191f78f1791b62a463234c3fde5fa924e3f8d1a845a65324" exitCode=0 Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.924821 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ccc4-account-create-pl9j2" event={"ID":"dad9687f-030b-4379-ba71-965b512fadaa","Type":"ContainerDied","Data":"d66c64aa24974228191f78f1791b62a463234c3fde5fa924e3f8d1a845a65324"} Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.924852 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ccc4-account-create-pl9j2" event={"ID":"dad9687f-030b-4379-ba71-965b512fadaa","Type":"ContainerStarted","Data":"9323bce9aabf80e40f013eff63b9878f186115b0c21b52c0a51113ba33b2f7c5"} Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.932918 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wmvlp" Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.935146 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-wmvlp" event={"ID":"49009ba5-e7dd-408f-ad33-a9fa20d156a6","Type":"ContainerDied","Data":"1260590f2a3f85fb266912e0bf32b87fdef4fb65e13c46944e8d4052f5798e89"} Oct 04 10:51:17 crc kubenswrapper[5025]: I1004 10:51:17.935196 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1260590f2a3f85fb266912e0bf32b87fdef4fb65e13c46944e8d4052f5798e89" Oct 04 10:51:17 crc kubenswrapper[5025]: W1004 10:51:17.984209 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod083601ee_774b_4e6d_a8e0_debebeb9011a.slice/crio-1fdcefd244b1ef558874f361cc3fa8fe4daf491ba3514e6007f28029c671b69b WatchSource:0}: Error finding container 1fdcefd244b1ef558874f361cc3fa8fe4daf491ba3514e6007f28029c671b69b: Status 404 returned error can't find the container with id 1fdcefd244b1ef558874f361cc3fa8fe4daf491ba3514e6007f28029c671b69b Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.037367 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jcfxn-config-4bczn"] Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.040173 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.045084 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.057951 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jcfxn-config-4bczn"] Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.150907 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdh2b\" (UniqueName: \"kubernetes.io/projected/c90cd925-d1bc-408f-b290-edf48d5cd54f-kube-api-access-fdh2b\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.151259 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run-ovn\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.151296 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-additional-scripts\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.151322 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-log-ovn\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.151351 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-scripts\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.151403 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.252620 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-scripts\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.252722 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.252782 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdh2b\" (UniqueName: \"kubernetes.io/projected/c90cd925-d1bc-408f-b290-edf48d5cd54f-kube-api-access-fdh2b\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.252839 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run-ovn\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.252877 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-additional-scripts\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.252905 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-log-ovn\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.253303 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-log-ovn\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.255435 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-scripts\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.255507 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.255766 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run-ovn\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.256391 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-additional-scripts\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.274218 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdh2b\" (UniqueName: \"kubernetes.io/projected/c90cd925-d1bc-408f-b290-edf48d5cd54f-kube-api-access-fdh2b\") pod \"ovn-controller-jcfxn-config-4bczn\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.363526 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.556192 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-frhrp"] Oct 04 10:51:18 crc kubenswrapper[5025]: W1004 10:51:18.581120 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d3aea52_35ad_49e4_93fb_465cd111f845.slice/crio-ab7d8bd00691ce068eb7c7ab6508b8d56a009a3cccad3ceafbb2e51ac8885c80 WatchSource:0}: Error finding container ab7d8bd00691ce068eb7c7ab6508b8d56a009a3cccad3ceafbb2e51ac8885c80: Status 404 returned error can't find the container with id ab7d8bd00691ce068eb7c7ab6508b8d56a009a3cccad3ceafbb2e51ac8885c80 Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.816110 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jcfxn-config-4bczn"] Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.940440 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"1145de6bbc5ed0cce04df8af6c009af1382ef214bf232e638c6fd1a3669d5118"} Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.940808 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"aca5db58e2caa71a849a6b4dcfbca078d7f2961edfe1203824c51e583950f372"} Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.940824 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"fc18498f31aacfc9c1100dcbc73e3f77f7f20dca71158de26dcd10bb3afd6d77"} Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.940833 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"e1cb1189f612c4c73a9180fea4e53ba6e51ae451e0279458f4baf44373e429f4"} Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.942353 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jcfxn-config-4bczn" event={"ID":"c90cd925-d1bc-408f-b290-edf48d5cd54f","Type":"ContainerStarted","Data":"0d41beca3ca38684441f7e123060cfb5d80e9ea82d525648098408412b6edc79"} Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.943891 5025 generic.go:334] "Generic (PLEG): container finished" podID="083601ee-774b-4e6d-a8e0-debebeb9011a" containerID="3b94de62d89966ce710fcb794f59a3425d495d60754a1e0b4fcced534c0223d4" exitCode=0 Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.943969 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9b7d-account-create-q2442" event={"ID":"083601ee-774b-4e6d-a8e0-debebeb9011a","Type":"ContainerDied","Data":"3b94de62d89966ce710fcb794f59a3425d495d60754a1e0b4fcced534c0223d4"} Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.943995 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9b7d-account-create-q2442" event={"ID":"083601ee-774b-4e6d-a8e0-debebeb9011a","Type":"ContainerStarted","Data":"1fdcefd244b1ef558874f361cc3fa8fe4daf491ba3514e6007f28029c671b69b"} Oct 04 10:51:18 crc kubenswrapper[5025]: I1004 10:51:18.945068 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-frhrp" event={"ID":"4d3aea52-35ad-49e4-93fb-465cd111f845","Type":"ContainerStarted","Data":"ab7d8bd00691ce068eb7c7ab6508b8d56a009a3cccad3ceafbb2e51ac8885c80"} Oct 04 10:51:19 crc kubenswrapper[5025]: I1004 10:51:19.208780 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ccc4-account-create-pl9j2" Oct 04 10:51:19 crc kubenswrapper[5025]: I1004 10:51:19.276114 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkkmv\" (UniqueName: \"kubernetes.io/projected/dad9687f-030b-4379-ba71-965b512fadaa-kube-api-access-rkkmv\") pod \"dad9687f-030b-4379-ba71-965b512fadaa\" (UID: \"dad9687f-030b-4379-ba71-965b512fadaa\") " Oct 04 10:51:19 crc kubenswrapper[5025]: I1004 10:51:19.283637 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dad9687f-030b-4379-ba71-965b512fadaa-kube-api-access-rkkmv" (OuterVolumeSpecName: "kube-api-access-rkkmv") pod "dad9687f-030b-4379-ba71-965b512fadaa" (UID: "dad9687f-030b-4379-ba71-965b512fadaa"). InnerVolumeSpecName "kube-api-access-rkkmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:19 crc kubenswrapper[5025]: I1004 10:51:19.378462 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkkmv\" (UniqueName: \"kubernetes.io/projected/dad9687f-030b-4379-ba71-965b512fadaa-kube-api-access-rkkmv\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:19 crc kubenswrapper[5025]: I1004 10:51:19.961936 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"8f547d885ffe33ba5def8412f6bb8916c48455fd2a8d9adce604f2b3c77d389b"} Oct 04 10:51:19 crc kubenswrapper[5025]: I1004 10:51:19.963753 5025 generic.go:334] "Generic (PLEG): container finished" podID="c90cd925-d1bc-408f-b290-edf48d5cd54f" containerID="fc3f2fdb59c5ed52cd622fdb5948a52993f2678899ea2a126225ec9f0b242589" exitCode=0 Oct 04 10:51:19 crc kubenswrapper[5025]: I1004 10:51:19.963809 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jcfxn-config-4bczn" event={"ID":"c90cd925-d1bc-408f-b290-edf48d5cd54f","Type":"ContainerDied","Data":"fc3f2fdb59c5ed52cd622fdb5948a52993f2678899ea2a126225ec9f0b242589"} Oct 04 10:51:19 crc kubenswrapper[5025]: I1004 10:51:19.965285 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ccc4-account-create-pl9j2" Oct 04 10:51:19 crc kubenswrapper[5025]: I1004 10:51:19.965812 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ccc4-account-create-pl9j2" event={"ID":"dad9687f-030b-4379-ba71-965b512fadaa","Type":"ContainerDied","Data":"9323bce9aabf80e40f013eff63b9878f186115b0c21b52c0a51113ba33b2f7c5"} Oct 04 10:51:19 crc kubenswrapper[5025]: I1004 10:51:19.965832 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9323bce9aabf80e40f013eff63b9878f186115b0c21b52c0a51113ba33b2f7c5" Oct 04 10:51:20 crc kubenswrapper[5025]: I1004 10:51:20.275455 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9b7d-account-create-q2442" Oct 04 10:51:20 crc kubenswrapper[5025]: I1004 10:51:20.394582 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m7jc\" (UniqueName: \"kubernetes.io/projected/083601ee-774b-4e6d-a8e0-debebeb9011a-kube-api-access-8m7jc\") pod \"083601ee-774b-4e6d-a8e0-debebeb9011a\" (UID: \"083601ee-774b-4e6d-a8e0-debebeb9011a\") " Oct 04 10:51:20 crc kubenswrapper[5025]: I1004 10:51:20.399629 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/083601ee-774b-4e6d-a8e0-debebeb9011a-kube-api-access-8m7jc" (OuterVolumeSpecName: "kube-api-access-8m7jc") pod "083601ee-774b-4e6d-a8e0-debebeb9011a" (UID: "083601ee-774b-4e6d-a8e0-debebeb9011a"). InnerVolumeSpecName "kube-api-access-8m7jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:20 crc kubenswrapper[5025]: I1004 10:51:20.496559 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m7jc\" (UniqueName: \"kubernetes.io/projected/083601ee-774b-4e6d-a8e0-debebeb9011a-kube-api-access-8m7jc\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:20 crc kubenswrapper[5025]: I1004 10:51:20.981166 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"487f1595d87ddc5e55726205a8f6b714e52ffc8387395b9c305b392020a65b3b"} Oct 04 10:51:20 crc kubenswrapper[5025]: I1004 10:51:20.983586 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-9b7d-account-create-q2442" event={"ID":"083601ee-774b-4e6d-a8e0-debebeb9011a","Type":"ContainerDied","Data":"1fdcefd244b1ef558874f361cc3fa8fe4daf491ba3514e6007f28029c671b69b"} Oct 04 10:51:20 crc kubenswrapper[5025]: I1004 10:51:20.983630 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fdcefd244b1ef558874f361cc3fa8fe4daf491ba3514e6007f28029c671b69b" Oct 04 10:51:20 crc kubenswrapper[5025]: I1004 10:51:20.983695 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-9b7d-account-create-q2442" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.274763 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.411873 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run\") pod \"c90cd925-d1bc-408f-b290-edf48d5cd54f\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.411927 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-additional-scripts\") pod \"c90cd925-d1bc-408f-b290-edf48d5cd54f\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.411979 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdh2b\" (UniqueName: \"kubernetes.io/projected/c90cd925-d1bc-408f-b290-edf48d5cd54f-kube-api-access-fdh2b\") pod \"c90cd925-d1bc-408f-b290-edf48d5cd54f\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.412003 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run" (OuterVolumeSpecName: "var-run") pod "c90cd925-d1bc-408f-b290-edf48d5cd54f" (UID: "c90cd925-d1bc-408f-b290-edf48d5cd54f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.412077 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run-ovn\") pod \"c90cd925-d1bc-408f-b290-edf48d5cd54f\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.412118 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-scripts\") pod \"c90cd925-d1bc-408f-b290-edf48d5cd54f\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.412155 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-log-ovn\") pod \"c90cd925-d1bc-408f-b290-edf48d5cd54f\" (UID: \"c90cd925-d1bc-408f-b290-edf48d5cd54f\") " Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.412173 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c90cd925-d1bc-408f-b290-edf48d5cd54f" (UID: "c90cd925-d1bc-408f-b290-edf48d5cd54f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.412377 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c90cd925-d1bc-408f-b290-edf48d5cd54f" (UID: "c90cd925-d1bc-408f-b290-edf48d5cd54f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.412853 5025 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.412894 5025 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.412917 5025 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c90cd925-d1bc-408f-b290-edf48d5cd54f-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.413180 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c90cd925-d1bc-408f-b290-edf48d5cd54f" (UID: "c90cd925-d1bc-408f-b290-edf48d5cd54f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.413757 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-scripts" (OuterVolumeSpecName: "scripts") pod "c90cd925-d1bc-408f-b290-edf48d5cd54f" (UID: "c90cd925-d1bc-408f-b290-edf48d5cd54f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.417934 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c90cd925-d1bc-408f-b290-edf48d5cd54f-kube-api-access-fdh2b" (OuterVolumeSpecName: "kube-api-access-fdh2b") pod "c90cd925-d1bc-408f-b290-edf48d5cd54f" (UID: "c90cd925-d1bc-408f-b290-edf48d5cd54f"). InnerVolumeSpecName "kube-api-access-fdh2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.514574 5025 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.514605 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdh2b\" (UniqueName: \"kubernetes.io/projected/c90cd925-d1bc-408f-b290-edf48d5cd54f-kube-api-access-fdh2b\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:21 crc kubenswrapper[5025]: I1004 10:51:21.514616 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c90cd925-d1bc-408f-b290-edf48d5cd54f-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.001105 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"4d8c0d9edfbbf81ef26a161a54023555eb8b873063fa769c0768ad733ac9cbb7"} Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.001155 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"0d4840d28afe1b56a23ccd69300c93c70b353e53eceea3d1e1a4233983886923"} Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.003175 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jcfxn-config-4bczn" event={"ID":"c90cd925-d1bc-408f-b290-edf48d5cd54f","Type":"ContainerDied","Data":"0d41beca3ca38684441f7e123060cfb5d80e9ea82d525648098408412b6edc79"} Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.003203 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d41beca3ca38684441f7e123060cfb5d80e9ea82d525648098408412b6edc79" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.003252 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jcfxn-config-4bczn" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.360743 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jcfxn-config-4bczn"] Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.374634 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jcfxn-config-4bczn"] Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.425998 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c90cd925-d1bc-408f-b290-edf48d5cd54f" path="/var/lib/kubelet/pods/c90cd925-d1bc-408f-b290-edf48d5cd54f/volumes" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.488587 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jcfxn-config-8sh6r"] Oct 04 10:51:22 crc kubenswrapper[5025]: E1004 10:51:22.488927 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dad9687f-030b-4379-ba71-965b512fadaa" containerName="mariadb-account-create" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.488946 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="dad9687f-030b-4379-ba71-965b512fadaa" containerName="mariadb-account-create" Oct 04 10:51:22 crc kubenswrapper[5025]: E1004 10:51:22.488962 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c90cd925-d1bc-408f-b290-edf48d5cd54f" containerName="ovn-config" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.488969 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="c90cd925-d1bc-408f-b290-edf48d5cd54f" containerName="ovn-config" Oct 04 10:51:22 crc kubenswrapper[5025]: E1004 10:51:22.488980 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="083601ee-774b-4e6d-a8e0-debebeb9011a" containerName="mariadb-account-create" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.488989 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="083601ee-774b-4e6d-a8e0-debebeb9011a" containerName="mariadb-account-create" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.489178 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="083601ee-774b-4e6d-a8e0-debebeb9011a" containerName="mariadb-account-create" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.489207 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="c90cd925-d1bc-408f-b290-edf48d5cd54f" containerName="ovn-config" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.489223 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="dad9687f-030b-4379-ba71-965b512fadaa" containerName="mariadb-account-create" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.490458 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.494438 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.502966 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jcfxn-config-8sh6r"] Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.648853 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.649217 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-log-ovn\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.649258 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwz92\" (UniqueName: \"kubernetes.io/projected/d4670832-4695-47d8-a8ea-a0add6019bb6-kube-api-access-wwz92\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.649294 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run-ovn\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.649322 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-scripts\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.649364 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-additional-scripts\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.750902 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-additional-scripts\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.750997 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.751033 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-log-ovn\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.751063 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwz92\" (UniqueName: \"kubernetes.io/projected/d4670832-4695-47d8-a8ea-a0add6019bb6-kube-api-access-wwz92\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.751096 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run-ovn\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.751118 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-scripts\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.753174 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-scripts\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.753624 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-additional-scripts\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.753896 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.753940 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-log-ovn\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.754227 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run-ovn\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.817550 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwz92\" (UniqueName: \"kubernetes.io/projected/d4670832-4695-47d8-a8ea-a0add6019bb6-kube-api-access-wwz92\") pod \"ovn-controller-jcfxn-config-8sh6r\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:22 crc kubenswrapper[5025]: I1004 10:51:22.904184 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jcfxn" Oct 04 10:51:23 crc kubenswrapper[5025]: I1004 10:51:23.026105 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"c40299777d003278e3aa771d7c723221510ef568bbfd7957a58217385c614e61"} Oct 04 10:51:23 crc kubenswrapper[5025]: I1004 10:51:23.026677 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"4f5ae42da8579276e7049f2be8c2aa9b5a1373630c182bbddea6b4b0ac284b44"} Oct 04 10:51:23 crc kubenswrapper[5025]: I1004 10:51:23.026714 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"a563dac5e62887eb0f0e07007235ef97d34b3343759fc2d8da841c7e370f815e"} Oct 04 10:51:23 crc kubenswrapper[5025]: I1004 10:51:23.114712 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:23 crc kubenswrapper[5025]: I1004 10:51:23.597080 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 10:51:23 crc kubenswrapper[5025]: W1004 10:51:23.645814 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4670832_4695_47d8_a8ea_a0add6019bb6.slice/crio-e0a1c643fe96c065d09b8afa6a7b90ab04328e2de4598e9c72758cc60e3fa08c WatchSource:0}: Error finding container e0a1c643fe96c065d09b8afa6a7b90ab04328e2de4598e9c72758cc60e3fa08c: Status 404 returned error can't find the container with id e0a1c643fe96c065d09b8afa6a7b90ab04328e2de4598e9c72758cc60e3fa08c Oct 04 10:51:23 crc kubenswrapper[5025]: I1004 10:51:23.658339 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jcfxn-config-8sh6r"] Oct 04 10:51:23 crc kubenswrapper[5025]: I1004 10:51:23.913220 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:51:23 crc kubenswrapper[5025]: I1004 10:51:23.959094 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-22h8p"] Oct 04 10:51:23 crc kubenswrapper[5025]: I1004 10:51:23.960340 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-22h8p" Oct 04 10:51:23 crc kubenswrapper[5025]: I1004 10:51:23.967277 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-22h8p"] Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.037902 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-bffxf"] Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.039505 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bffxf" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.058270 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bffxf"] Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.064922 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"6a1d4de31fee41d61f331e3564c776766158e9beb47846d11a289ab138c7c87c"} Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.064981 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"2cc34d6bfa0ee8f912bf61c44cc9b07338c0d8caf533f8978f2114b87c443395"} Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.064995 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"23808863d20800db9b97efd941c6667b90e967a575778077551f4b468ff4e341"} Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.065008 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"369e99d7-ddff-4bfb-9bea-a382aa69eb3e","Type":"ContainerStarted","Data":"e012c90b817ea97cee84458b5db3d5d0d91711c30893716fe1f7b37e0e395cac"} Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.074562 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jcfxn-config-8sh6r" event={"ID":"d4670832-4695-47d8-a8ea-a0add6019bb6","Type":"ContainerStarted","Data":"e0a1c643fe96c065d09b8afa6a7b90ab04328e2de4598e9c72758cc60e3fa08c"} Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.083444 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbmj8\" (UniqueName: \"kubernetes.io/projected/26b3df34-d44c-4f4a-888d-73413cd3a358-kube-api-access-pbmj8\") pod \"cinder-db-create-22h8p\" (UID: \"26b3df34-d44c-4f4a-888d-73413cd3a358\") " pod="openstack/cinder-db-create-22h8p" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.083540 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fcxr\" (UniqueName: \"kubernetes.io/projected/31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05-kube-api-access-9fcxr\") pod \"barbican-db-create-bffxf\" (UID: \"31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05\") " pod="openstack/barbican-db-create-bffxf" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.129777 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.841005167 podStartE2EDuration="25.129759469s" podCreationTimestamp="2025-10-04 10:50:59 +0000 UTC" firstStartedPulling="2025-10-04 10:51:16.800714221 +0000 UTC m=+1005.225681111" lastFinishedPulling="2025-10-04 10:51:22.089468533 +0000 UTC m=+1010.514435413" observedRunningTime="2025-10-04 10:51:24.125464324 +0000 UTC m=+1012.550431224" watchObservedRunningTime="2025-10-04 10:51:24.129759469 +0000 UTC m=+1012.554726349" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.187511 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fcxr\" (UniqueName: \"kubernetes.io/projected/31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05-kube-api-access-9fcxr\") pod \"barbican-db-create-bffxf\" (UID: \"31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05\") " pod="openstack/barbican-db-create-bffxf" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.187756 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbmj8\" (UniqueName: \"kubernetes.io/projected/26b3df34-d44c-4f4a-888d-73413cd3a358-kube-api-access-pbmj8\") pod \"cinder-db-create-22h8p\" (UID: \"26b3df34-d44c-4f4a-888d-73413cd3a358\") " pod="openstack/cinder-db-create-22h8p" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.207136 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbmj8\" (UniqueName: \"kubernetes.io/projected/26b3df34-d44c-4f4a-888d-73413cd3a358-kube-api-access-pbmj8\") pod \"cinder-db-create-22h8p\" (UID: \"26b3df34-d44c-4f4a-888d-73413cd3a358\") " pod="openstack/cinder-db-create-22h8p" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.211550 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fcxr\" (UniqueName: \"kubernetes.io/projected/31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05-kube-api-access-9fcxr\") pod \"barbican-db-create-bffxf\" (UID: \"31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05\") " pod="openstack/barbican-db-create-bffxf" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.287568 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-ctpv7"] Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.288608 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.294741 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.294916 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.295104 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9d9vx" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.301310 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ctpv7"] Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.301453 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.372960 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-22h8p" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.374316 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-xlm27"] Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.377547 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xlm27" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.390531 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-combined-ca-bundle\") pod \"keystone-db-sync-ctpv7\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.390865 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8vsc\" (UniqueName: \"kubernetes.io/projected/1270a45a-df30-4138-8270-78174433f9cb-kube-api-access-v8vsc\") pod \"keystone-db-sync-ctpv7\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.391124 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-config-data\") pod \"keystone-db-sync-ctpv7\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.402131 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xlm27"] Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.402969 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bffxf" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.452374 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrdkc"] Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.453595 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.469967 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.473601 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrdkc"] Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.498548 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-config\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.498596 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f95qq\" (UniqueName: \"kubernetes.io/projected/021282fc-71aa-4205-9626-24bd450b837e-kube-api-access-f95qq\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.498667 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.498704 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8vsc\" (UniqueName: \"kubernetes.io/projected/1270a45a-df30-4138-8270-78174433f9cb-kube-api-access-v8vsc\") pod \"keystone-db-sync-ctpv7\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.498741 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-config-data\") pod \"keystone-db-sync-ctpv7\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.498775 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.498833 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drmnq\" (UniqueName: \"kubernetes.io/projected/98f93315-6097-4365-9972-03e9559f8c32-kube-api-access-drmnq\") pod \"neutron-db-create-xlm27\" (UID: \"98f93315-6097-4365-9972-03e9559f8c32\") " pod="openstack/neutron-db-create-xlm27" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.498867 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-combined-ca-bundle\") pod \"keystone-db-sync-ctpv7\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.498890 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.498927 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.511099 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-config-data\") pod \"keystone-db-sync-ctpv7\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.516706 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-combined-ca-bundle\") pod \"keystone-db-sync-ctpv7\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.525188 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8vsc\" (UniqueName: \"kubernetes.io/projected/1270a45a-df30-4138-8270-78174433f9cb-kube-api-access-v8vsc\") pod \"keystone-db-sync-ctpv7\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.600204 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drmnq\" (UniqueName: \"kubernetes.io/projected/98f93315-6097-4365-9972-03e9559f8c32-kube-api-access-drmnq\") pod \"neutron-db-create-xlm27\" (UID: \"98f93315-6097-4365-9972-03e9559f8c32\") " pod="openstack/neutron-db-create-xlm27" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.600365 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.600411 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.600467 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-config\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.600483 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f95qq\" (UniqueName: \"kubernetes.io/projected/021282fc-71aa-4205-9626-24bd450b837e-kube-api-access-f95qq\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.600522 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.600566 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.602170 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-config\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.602392 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.603262 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.603952 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.608934 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.617969 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f95qq\" (UniqueName: \"kubernetes.io/projected/021282fc-71aa-4205-9626-24bd450b837e-kube-api-access-f95qq\") pod \"dnsmasq-dns-6d5b6d6b67-nrdkc\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.618809 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drmnq\" (UniqueName: \"kubernetes.io/projected/98f93315-6097-4365-9972-03e9559f8c32-kube-api-access-drmnq\") pod \"neutron-db-create-xlm27\" (UID: \"98f93315-6097-4365-9972-03e9559f8c32\") " pod="openstack/neutron-db-create-xlm27" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.658288 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.697435 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xlm27" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.916621 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:24 crc kubenswrapper[5025]: I1004 10:51:24.934131 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-22h8p"] Oct 04 10:51:24 crc kubenswrapper[5025]: W1004 10:51:24.971070 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26b3df34_d44c_4f4a_888d_73413cd3a358.slice/crio-9bbae77c294ae21bd35c3c0efbbb34a5b45cf8b39885180fcaa63a40cbe434c1 WatchSource:0}: Error finding container 9bbae77c294ae21bd35c3c0efbbb34a5b45cf8b39885180fcaa63a40cbe434c1: Status 404 returned error can't find the container with id 9bbae77c294ae21bd35c3c0efbbb34a5b45cf8b39885180fcaa63a40cbe434c1 Oct 04 10:51:25 crc kubenswrapper[5025]: I1004 10:51:25.011350 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-bffxf"] Oct 04 10:51:25 crc kubenswrapper[5025]: I1004 10:51:25.090448 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-22h8p" event={"ID":"26b3df34-d44c-4f4a-888d-73413cd3a358","Type":"ContainerStarted","Data":"9bbae77c294ae21bd35c3c0efbbb34a5b45cf8b39885180fcaa63a40cbe434c1"} Oct 04 10:51:25 crc kubenswrapper[5025]: I1004 10:51:25.091776 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bffxf" event={"ID":"31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05","Type":"ContainerStarted","Data":"3792baaef1c90dbd7cf6eccea146f373d66b2eb423ea72648d7c9469fd7e3a3b"} Oct 04 10:51:25 crc kubenswrapper[5025]: I1004 10:51:25.093825 5025 generic.go:334] "Generic (PLEG): container finished" podID="d4670832-4695-47d8-a8ea-a0add6019bb6" containerID="4ad132bcb6db1791f620f0fb33be8a0e5e1299b62a58e7914f83bfc9d68d491e" exitCode=0 Oct 04 10:51:25 crc kubenswrapper[5025]: I1004 10:51:25.094155 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jcfxn-config-8sh6r" event={"ID":"d4670832-4695-47d8-a8ea-a0add6019bb6","Type":"ContainerDied","Data":"4ad132bcb6db1791f620f0fb33be8a0e5e1299b62a58e7914f83bfc9d68d491e"} Oct 04 10:51:25 crc kubenswrapper[5025]: I1004 10:51:25.182198 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xlm27"] Oct 04 10:51:25 crc kubenswrapper[5025]: I1004 10:51:25.251735 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ctpv7"] Oct 04 10:51:25 crc kubenswrapper[5025]: I1004 10:51:25.411632 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrdkc"] Oct 04 10:51:26 crc kubenswrapper[5025]: I1004 10:51:26.104457 5025 generic.go:334] "Generic (PLEG): container finished" podID="31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05" containerID="4d306b6d67206c37b584322b76a665a377d05d028e7cb1f5b991b7c1836e5239" exitCode=0 Oct 04 10:51:26 crc kubenswrapper[5025]: I1004 10:51:26.104546 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bffxf" event={"ID":"31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05","Type":"ContainerDied","Data":"4d306b6d67206c37b584322b76a665a377d05d028e7cb1f5b991b7c1836e5239"} Oct 04 10:51:26 crc kubenswrapper[5025]: I1004 10:51:26.105917 5025 generic.go:334] "Generic (PLEG): container finished" podID="26b3df34-d44c-4f4a-888d-73413cd3a358" containerID="e9eaba596fa46d38b6f26e181da1df105847685501383c957151836c04ef4d5d" exitCode=0 Oct 04 10:51:26 crc kubenswrapper[5025]: I1004 10:51:26.105989 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-22h8p" event={"ID":"26b3df34-d44c-4f4a-888d-73413cd3a358","Type":"ContainerDied","Data":"e9eaba596fa46d38b6f26e181da1df105847685501383c957151836c04ef4d5d"} Oct 04 10:51:31 crc kubenswrapper[5025]: W1004 10:51:31.902989 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98f93315_6097_4365_9972_03e9559f8c32.slice/crio-e11592081f9e85c9b297684bf6219697a5f136170902e237e9efa430b418b38f WatchSource:0}: Error finding container e11592081f9e85c9b297684bf6219697a5f136170902e237e9efa430b418b38f: Status 404 returned error can't find the container with id e11592081f9e85c9b297684bf6219697a5f136170902e237e9efa430b418b38f Oct 04 10:51:31 crc kubenswrapper[5025]: W1004 10:51:31.905948 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1270a45a_df30_4138_8270_78174433f9cb.slice/crio-bac07a0bfb19984d0c026ce7af78618628555a93e776211eb7b8955c571475a5 WatchSource:0}: Error finding container bac07a0bfb19984d0c026ce7af78618628555a93e776211eb7b8955c571475a5: Status 404 returned error can't find the container with id bac07a0bfb19984d0c026ce7af78618628555a93e776211eb7b8955c571475a5 Oct 04 10:51:31 crc kubenswrapper[5025]: W1004 10:51:31.907029 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod021282fc_71aa_4205_9626_24bd450b837e.slice/crio-0dc35a110c597c215407d966dcb34fef1b56bfc78e1fa55941a94dde7cb367d3 WatchSource:0}: Error finding container 0dc35a110c597c215407d966dcb34fef1b56bfc78e1fa55941a94dde7cb367d3: Status 404 returned error can't find the container with id 0dc35a110c597c215407d966dcb34fef1b56bfc78e1fa55941a94dde7cb367d3 Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.162744 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" event={"ID":"021282fc-71aa-4205-9626-24bd450b837e","Type":"ContainerStarted","Data":"0dc35a110c597c215407d966dcb34fef1b56bfc78e1fa55941a94dde7cb367d3"} Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.165434 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-bffxf" event={"ID":"31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05","Type":"ContainerDied","Data":"3792baaef1c90dbd7cf6eccea146f373d66b2eb423ea72648d7c9469fd7e3a3b"} Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.165467 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3792baaef1c90dbd7cf6eccea146f373d66b2eb423ea72648d7c9469fd7e3a3b" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.168055 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jcfxn-config-8sh6r" event={"ID":"d4670832-4695-47d8-a8ea-a0add6019bb6","Type":"ContainerDied","Data":"e0a1c643fe96c065d09b8afa6a7b90ab04328e2de4598e9c72758cc60e3fa08c"} Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.168086 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0a1c643fe96c065d09b8afa6a7b90ab04328e2de4598e9c72758cc60e3fa08c" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.169483 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xlm27" event={"ID":"98f93315-6097-4365-9972-03e9559f8c32","Type":"ContainerStarted","Data":"e11592081f9e85c9b297684bf6219697a5f136170902e237e9efa430b418b38f"} Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.170634 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ctpv7" event={"ID":"1270a45a-df30-4138-8270-78174433f9cb","Type":"ContainerStarted","Data":"bac07a0bfb19984d0c026ce7af78618628555a93e776211eb7b8955c571475a5"} Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.171998 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-22h8p" event={"ID":"26b3df34-d44c-4f4a-888d-73413cd3a358","Type":"ContainerDied","Data":"9bbae77c294ae21bd35c3c0efbbb34a5b45cf8b39885180fcaa63a40cbe434c1"} Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.172031 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bbae77c294ae21bd35c3c0efbbb34a5b45cf8b39885180fcaa63a40cbe434c1" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.211114 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.250713 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bffxf" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.277198 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-22h8p" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.329492 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-additional-scripts\") pod \"d4670832-4695-47d8-a8ea-a0add6019bb6\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.329537 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-log-ovn\") pod \"d4670832-4695-47d8-a8ea-a0add6019bb6\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.329573 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbmj8\" (UniqueName: \"kubernetes.io/projected/26b3df34-d44c-4f4a-888d-73413cd3a358-kube-api-access-pbmj8\") pod \"26b3df34-d44c-4f4a-888d-73413cd3a358\" (UID: \"26b3df34-d44c-4f4a-888d-73413cd3a358\") " Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.329644 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-scripts\") pod \"d4670832-4695-47d8-a8ea-a0add6019bb6\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.329697 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run-ovn\") pod \"d4670832-4695-47d8-a8ea-a0add6019bb6\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.329817 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fcxr\" (UniqueName: \"kubernetes.io/projected/31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05-kube-api-access-9fcxr\") pod \"31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05\" (UID: \"31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05\") " Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.329867 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run\") pod \"d4670832-4695-47d8-a8ea-a0add6019bb6\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.329891 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwz92\" (UniqueName: \"kubernetes.io/projected/d4670832-4695-47d8-a8ea-a0add6019bb6-kube-api-access-wwz92\") pod \"d4670832-4695-47d8-a8ea-a0add6019bb6\" (UID: \"d4670832-4695-47d8-a8ea-a0add6019bb6\") " Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.330796 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run" (OuterVolumeSpecName: "var-run") pod "d4670832-4695-47d8-a8ea-a0add6019bb6" (UID: "d4670832-4695-47d8-a8ea-a0add6019bb6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.330859 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d4670832-4695-47d8-a8ea-a0add6019bb6" (UID: "d4670832-4695-47d8-a8ea-a0add6019bb6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.330876 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d4670832-4695-47d8-a8ea-a0add6019bb6" (UID: "d4670832-4695-47d8-a8ea-a0add6019bb6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.331297 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d4670832-4695-47d8-a8ea-a0add6019bb6" (UID: "d4670832-4695-47d8-a8ea-a0add6019bb6"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.331643 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-scripts" (OuterVolumeSpecName: "scripts") pod "d4670832-4695-47d8-a8ea-a0add6019bb6" (UID: "d4670832-4695-47d8-a8ea-a0add6019bb6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.336215 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05-kube-api-access-9fcxr" (OuterVolumeSpecName: "kube-api-access-9fcxr") pod "31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05" (UID: "31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05"). InnerVolumeSpecName "kube-api-access-9fcxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.336514 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26b3df34-d44c-4f4a-888d-73413cd3a358-kube-api-access-pbmj8" (OuterVolumeSpecName: "kube-api-access-pbmj8") pod "26b3df34-d44c-4f4a-888d-73413cd3a358" (UID: "26b3df34-d44c-4f4a-888d-73413cd3a358"). InnerVolumeSpecName "kube-api-access-pbmj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.338369 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4670832-4695-47d8-a8ea-a0add6019bb6-kube-api-access-wwz92" (OuterVolumeSpecName: "kube-api-access-wwz92") pod "d4670832-4695-47d8-a8ea-a0add6019bb6" (UID: "d4670832-4695-47d8-a8ea-a0add6019bb6"). InnerVolumeSpecName "kube-api-access-wwz92". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.431275 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.431301 5025 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.431311 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fcxr\" (UniqueName: \"kubernetes.io/projected/31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05-kube-api-access-9fcxr\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.431343 5025 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-run\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.431351 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwz92\" (UniqueName: \"kubernetes.io/projected/d4670832-4695-47d8-a8ea-a0add6019bb6-kube-api-access-wwz92\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.431360 5025 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d4670832-4695-47d8-a8ea-a0add6019bb6-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.431368 5025 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d4670832-4695-47d8-a8ea-a0add6019bb6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:32 crc kubenswrapper[5025]: I1004 10:51:32.431376 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbmj8\" (UniqueName: \"kubernetes.io/projected/26b3df34-d44c-4f4a-888d-73413cd3a358-kube-api-access-pbmj8\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.185209 5025 generic.go:334] "Generic (PLEG): container finished" podID="98f93315-6097-4365-9972-03e9559f8c32" containerID="aa7acc7ce1d3f4d7db14c73df69839aa8cac991c3589c934777f00ef5e75e4bf" exitCode=0 Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.185287 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xlm27" event={"ID":"98f93315-6097-4365-9972-03e9559f8c32","Type":"ContainerDied","Data":"aa7acc7ce1d3f4d7db14c73df69839aa8cac991c3589c934777f00ef5e75e4bf"} Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.191867 5025 generic.go:334] "Generic (PLEG): container finished" podID="021282fc-71aa-4205-9626-24bd450b837e" containerID="0bf38c469326095aa149c09982e2dfbee19aa502fe7dbda60dec10caeefe4abe" exitCode=0 Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.191951 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" event={"ID":"021282fc-71aa-4205-9626-24bd450b837e","Type":"ContainerDied","Data":"0bf38c469326095aa149c09982e2dfbee19aa502fe7dbda60dec10caeefe4abe"} Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.194080 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-frhrp" event={"ID":"4d3aea52-35ad-49e4-93fb-465cd111f845","Type":"ContainerStarted","Data":"22c2e8b34fd6e772d90658dccb31c3f9b4484035315f76d6b01e48d3d56d0157"} Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.194132 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-22h8p" Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.194138 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-bffxf" Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.194295 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jcfxn-config-8sh6r" Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.227109 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-frhrp" podStartSLOduration=2.784028147 podStartE2EDuration="16.227090515s" podCreationTimestamp="2025-10-04 10:51:17 +0000 UTC" firstStartedPulling="2025-10-04 10:51:18.585796709 +0000 UTC m=+1007.010763589" lastFinishedPulling="2025-10-04 10:51:32.028859067 +0000 UTC m=+1020.453825957" observedRunningTime="2025-10-04 10:51:33.223310565 +0000 UTC m=+1021.648277465" watchObservedRunningTime="2025-10-04 10:51:33.227090515 +0000 UTC m=+1021.652057395" Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.307259 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jcfxn-config-8sh6r"] Oct 04 10:51:33 crc kubenswrapper[5025]: I1004 10:51:33.315996 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jcfxn-config-8sh6r"] Oct 04 10:51:34 crc kubenswrapper[5025]: I1004 10:51:34.205287 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" event={"ID":"021282fc-71aa-4205-9626-24bd450b837e","Type":"ContainerStarted","Data":"9c9eb64ba60da9dc8eae7140cb2610c6219b736b7f89ee5e1c7721f1bad4e904"} Oct 04 10:51:34 crc kubenswrapper[5025]: I1004 10:51:34.225296 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" podStartSLOduration=10.225254326 podStartE2EDuration="10.225254326s" podCreationTimestamp="2025-10-04 10:51:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:51:34.222443144 +0000 UTC m=+1022.647410024" watchObservedRunningTime="2025-10-04 10:51:34.225254326 +0000 UTC m=+1022.650221206" Oct 04 10:51:34 crc kubenswrapper[5025]: I1004 10:51:34.422166 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4670832-4695-47d8-a8ea-a0add6019bb6" path="/var/lib/kubelet/pods/d4670832-4695-47d8-a8ea-a0add6019bb6/volumes" Oct 04 10:51:34 crc kubenswrapper[5025]: I1004 10:51:34.919498 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:36 crc kubenswrapper[5025]: I1004 10:51:36.653068 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xlm27" Oct 04 10:51:36 crc kubenswrapper[5025]: I1004 10:51:36.727761 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drmnq\" (UniqueName: \"kubernetes.io/projected/98f93315-6097-4365-9972-03e9559f8c32-kube-api-access-drmnq\") pod \"98f93315-6097-4365-9972-03e9559f8c32\" (UID: \"98f93315-6097-4365-9972-03e9559f8c32\") " Oct 04 10:51:36 crc kubenswrapper[5025]: I1004 10:51:36.734451 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98f93315-6097-4365-9972-03e9559f8c32-kube-api-access-drmnq" (OuterVolumeSpecName: "kube-api-access-drmnq") pod "98f93315-6097-4365-9972-03e9559f8c32" (UID: "98f93315-6097-4365-9972-03e9559f8c32"). InnerVolumeSpecName "kube-api-access-drmnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:36 crc kubenswrapper[5025]: I1004 10:51:36.830583 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drmnq\" (UniqueName: \"kubernetes.io/projected/98f93315-6097-4365-9972-03e9559f8c32-kube-api-access-drmnq\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:37 crc kubenswrapper[5025]: I1004 10:51:37.245558 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xlm27" event={"ID":"98f93315-6097-4365-9972-03e9559f8c32","Type":"ContainerDied","Data":"e11592081f9e85c9b297684bf6219697a5f136170902e237e9efa430b418b38f"} Oct 04 10:51:37 crc kubenswrapper[5025]: I1004 10:51:37.245600 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e11592081f9e85c9b297684bf6219697a5f136170902e237e9efa430b418b38f" Oct 04 10:51:37 crc kubenswrapper[5025]: I1004 10:51:37.245995 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xlm27" Oct 04 10:51:37 crc kubenswrapper[5025]: E1004 10:51:37.357951 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98f93315_6097_4365_9972_03e9559f8c32.slice\": RecentStats: unable to find data in memory cache]" Oct 04 10:51:39 crc kubenswrapper[5025]: I1004 10:51:39.919771 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:51:39 crc kubenswrapper[5025]: I1004 10:51:39.989807 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-294dc"] Oct 04 10:51:39 crc kubenswrapper[5025]: I1004 10:51:39.990201 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" podUID="afc4132c-4f96-4898-b18c-4bc79168f8af" containerName="dnsmasq-dns" containerID="cri-o://948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0" gracePeriod=10 Oct 04 10:51:42 crc kubenswrapper[5025]: I1004 10:51:42.977761 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.053788 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-config\") pod \"afc4132c-4f96-4898-b18c-4bc79168f8af\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.053820 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-nb\") pod \"afc4132c-4f96-4898-b18c-4bc79168f8af\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.053844 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-sb\") pod \"afc4132c-4f96-4898-b18c-4bc79168f8af\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.053918 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-dns-svc\") pod \"afc4132c-4f96-4898-b18c-4bc79168f8af\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.054041 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cknh\" (UniqueName: \"kubernetes.io/projected/afc4132c-4f96-4898-b18c-4bc79168f8af-kube-api-access-8cknh\") pod \"afc4132c-4f96-4898-b18c-4bc79168f8af\" (UID: \"afc4132c-4f96-4898-b18c-4bc79168f8af\") " Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.067551 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc4132c-4f96-4898-b18c-4bc79168f8af-kube-api-access-8cknh" (OuterVolumeSpecName: "kube-api-access-8cknh") pod "afc4132c-4f96-4898-b18c-4bc79168f8af" (UID: "afc4132c-4f96-4898-b18c-4bc79168f8af"). InnerVolumeSpecName "kube-api-access-8cknh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.101004 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-config" (OuterVolumeSpecName: "config") pod "afc4132c-4f96-4898-b18c-4bc79168f8af" (UID: "afc4132c-4f96-4898-b18c-4bc79168f8af"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.103397 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "afc4132c-4f96-4898-b18c-4bc79168f8af" (UID: "afc4132c-4f96-4898-b18c-4bc79168f8af"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.103799 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "afc4132c-4f96-4898-b18c-4bc79168f8af" (UID: "afc4132c-4f96-4898-b18c-4bc79168f8af"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.106589 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "afc4132c-4f96-4898-b18c-4bc79168f8af" (UID: "afc4132c-4f96-4898-b18c-4bc79168f8af"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.155994 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cknh\" (UniqueName: \"kubernetes.io/projected/afc4132c-4f96-4898-b18c-4bc79168f8af-kube-api-access-8cknh\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.156040 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.156049 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.156058 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.156065 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/afc4132c-4f96-4898-b18c-4bc79168f8af-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.308659 5025 generic.go:334] "Generic (PLEG): container finished" podID="afc4132c-4f96-4898-b18c-4bc79168f8af" containerID="948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0" exitCode=0 Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.308736 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" event={"ID":"afc4132c-4f96-4898-b18c-4bc79168f8af","Type":"ContainerDied","Data":"948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0"} Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.308767 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" event={"ID":"afc4132c-4f96-4898-b18c-4bc79168f8af","Type":"ContainerDied","Data":"0997d5944e57972de5e4cd5138a41314ce4fb1332ec306afe35ba8c44e08f7b6"} Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.308793 5025 scope.go:117] "RemoveContainer" containerID="948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.308913 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-294dc" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.319513 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ctpv7" event={"ID":"1270a45a-df30-4138-8270-78174433f9cb","Type":"ContainerStarted","Data":"149b5de7d7e8cea0b473cf87d614843df1451b44ccf2ee3c07c16f2c2af9005e"} Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.345881 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-ctpv7" podStartSLOduration=8.659117167 podStartE2EDuration="19.345860852s" podCreationTimestamp="2025-10-04 10:51:24 +0000 UTC" firstStartedPulling="2025-10-04 10:51:31.958394848 +0000 UTC m=+1020.383361728" lastFinishedPulling="2025-10-04 10:51:42.645138533 +0000 UTC m=+1031.070105413" observedRunningTime="2025-10-04 10:51:43.339930788 +0000 UTC m=+1031.764897678" watchObservedRunningTime="2025-10-04 10:51:43.345860852 +0000 UTC m=+1031.770827752" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.362724 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-294dc"] Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.366935 5025 scope.go:117] "RemoveContainer" containerID="ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.371354 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-294dc"] Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.384715 5025 scope.go:117] "RemoveContainer" containerID="948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0" Oct 04 10:51:43 crc kubenswrapper[5025]: E1004 10:51:43.385755 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0\": container with ID starting with 948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0 not found: ID does not exist" containerID="948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.385822 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0"} err="failed to get container status \"948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0\": rpc error: code = NotFound desc = could not find container \"948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0\": container with ID starting with 948da2a4b6da63741e030c53c55a9c70b9989ae8c32f048bb929f509339ca0b0 not found: ID does not exist" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.385862 5025 scope.go:117] "RemoveContainer" containerID="ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5" Oct 04 10:51:43 crc kubenswrapper[5025]: E1004 10:51:43.386226 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5\": container with ID starting with ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5 not found: ID does not exist" containerID="ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.386272 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5"} err="failed to get container status \"ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5\": rpc error: code = NotFound desc = could not find container \"ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5\": container with ID starting with ccb7d6f16fdd3e420abe0a6fc8d286db34968dd1df63d68ecea585b7d4917cf5 not found: ID does not exist" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.962626 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c3e5-account-create-sxqrv"] Oct 04 10:51:43 crc kubenswrapper[5025]: E1004 10:51:43.963364 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc4132c-4f96-4898-b18c-4bc79168f8af" containerName="dnsmasq-dns" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.963479 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc4132c-4f96-4898-b18c-4bc79168f8af" containerName="dnsmasq-dns" Oct 04 10:51:43 crc kubenswrapper[5025]: E1004 10:51:43.963573 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc4132c-4f96-4898-b18c-4bc79168f8af" containerName="init" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.963663 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc4132c-4f96-4898-b18c-4bc79168f8af" containerName="init" Oct 04 10:51:43 crc kubenswrapper[5025]: E1004 10:51:43.963754 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4670832-4695-47d8-a8ea-a0add6019bb6" containerName="ovn-config" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.963841 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4670832-4695-47d8-a8ea-a0add6019bb6" containerName="ovn-config" Oct 04 10:51:43 crc kubenswrapper[5025]: E1004 10:51:43.963947 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98f93315-6097-4365-9972-03e9559f8c32" containerName="mariadb-database-create" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.964060 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="98f93315-6097-4365-9972-03e9559f8c32" containerName="mariadb-database-create" Oct 04 10:51:43 crc kubenswrapper[5025]: E1004 10:51:43.964212 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26b3df34-d44c-4f4a-888d-73413cd3a358" containerName="mariadb-database-create" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.964297 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="26b3df34-d44c-4f4a-888d-73413cd3a358" containerName="mariadb-database-create" Oct 04 10:51:43 crc kubenswrapper[5025]: E1004 10:51:43.964383 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05" containerName="mariadb-database-create" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.964468 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05" containerName="mariadb-database-create" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.964778 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4670832-4695-47d8-a8ea-a0add6019bb6" containerName="ovn-config" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.964905 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="26b3df34-d44c-4f4a-888d-73413cd3a358" containerName="mariadb-database-create" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.964994 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="98f93315-6097-4365-9972-03e9559f8c32" containerName="mariadb-database-create" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.965119 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05" containerName="mariadb-database-create" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.965218 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc4132c-4f96-4898-b18c-4bc79168f8af" containerName="dnsmasq-dns" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.966168 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c3e5-account-create-sxqrv" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.968481 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 04 10:51:43 crc kubenswrapper[5025]: I1004 10:51:43.971377 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c3e5-account-create-sxqrv"] Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.069440 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-1b98-account-create-c872x"] Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.070410 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b2qx\" (UniqueName: \"kubernetes.io/projected/4323bbf4-8801-411a-8fce-bd4a077180da-kube-api-access-2b2qx\") pod \"barbican-c3e5-account-create-sxqrv\" (UID: \"4323bbf4-8801-411a-8fce-bd4a077180da\") " pod="openstack/barbican-c3e5-account-create-sxqrv" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.070842 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1b98-account-create-c872x" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.074421 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.078935 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1b98-account-create-c872x"] Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.172298 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b2qx\" (UniqueName: \"kubernetes.io/projected/4323bbf4-8801-411a-8fce-bd4a077180da-kube-api-access-2b2qx\") pod \"barbican-c3e5-account-create-sxqrv\" (UID: \"4323bbf4-8801-411a-8fce-bd4a077180da\") " pod="openstack/barbican-c3e5-account-create-sxqrv" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.172405 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll55n\" (UniqueName: \"kubernetes.io/projected/298276d3-1df7-4934-88b3-cfcbfbc0fd51-kube-api-access-ll55n\") pod \"cinder-1b98-account-create-c872x\" (UID: \"298276d3-1df7-4934-88b3-cfcbfbc0fd51\") " pod="openstack/cinder-1b98-account-create-c872x" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.188196 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b2qx\" (UniqueName: \"kubernetes.io/projected/4323bbf4-8801-411a-8fce-bd4a077180da-kube-api-access-2b2qx\") pod \"barbican-c3e5-account-create-sxqrv\" (UID: \"4323bbf4-8801-411a-8fce-bd4a077180da\") " pod="openstack/barbican-c3e5-account-create-sxqrv" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.274178 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll55n\" (UniqueName: \"kubernetes.io/projected/298276d3-1df7-4934-88b3-cfcbfbc0fd51-kube-api-access-ll55n\") pod \"cinder-1b98-account-create-c872x\" (UID: \"298276d3-1df7-4934-88b3-cfcbfbc0fd51\") " pod="openstack/cinder-1b98-account-create-c872x" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.283812 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c3e5-account-create-sxqrv" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.292747 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll55n\" (UniqueName: \"kubernetes.io/projected/298276d3-1df7-4934-88b3-cfcbfbc0fd51-kube-api-access-ll55n\") pod \"cinder-1b98-account-create-c872x\" (UID: \"298276d3-1df7-4934-88b3-cfcbfbc0fd51\") " pod="openstack/cinder-1b98-account-create-c872x" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.366673 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c114-account-create-ttrlz"] Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.388660 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1b98-account-create-c872x" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.389437 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c114-account-create-ttrlz" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.392778 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.405538 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c114-account-create-ttrlz"] Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.427697 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc4132c-4f96-4898-b18c-4bc79168f8af" path="/var/lib/kubelet/pods/afc4132c-4f96-4898-b18c-4bc79168f8af/volumes" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.486072 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g74d9\" (UniqueName: \"kubernetes.io/projected/ce76a45b-1519-4a6e-9d7a-7a33bebadc13-kube-api-access-g74d9\") pod \"neutron-c114-account-create-ttrlz\" (UID: \"ce76a45b-1519-4a6e-9d7a-7a33bebadc13\") " pod="openstack/neutron-c114-account-create-ttrlz" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.588024 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g74d9\" (UniqueName: \"kubernetes.io/projected/ce76a45b-1519-4a6e-9d7a-7a33bebadc13-kube-api-access-g74d9\") pod \"neutron-c114-account-create-ttrlz\" (UID: \"ce76a45b-1519-4a6e-9d7a-7a33bebadc13\") " pod="openstack/neutron-c114-account-create-ttrlz" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.605212 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g74d9\" (UniqueName: \"kubernetes.io/projected/ce76a45b-1519-4a6e-9d7a-7a33bebadc13-kube-api-access-g74d9\") pod \"neutron-c114-account-create-ttrlz\" (UID: \"ce76a45b-1519-4a6e-9d7a-7a33bebadc13\") " pod="openstack/neutron-c114-account-create-ttrlz" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.714035 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.714128 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.714193 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.715270 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1fa2c044718960ddbde296231834fb7f5a1519308e0e7742cc723c16081ff583"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.715373 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://1fa2c044718960ddbde296231834fb7f5a1519308e0e7742cc723c16081ff583" gracePeriod=600 Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.760985 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c3e5-account-create-sxqrv"] Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.771864 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c114-account-create-ttrlz" Oct 04 10:51:44 crc kubenswrapper[5025]: W1004 10:51:44.773548 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4323bbf4_8801_411a_8fce_bd4a077180da.slice/crio-7e4c46378e2b20c9fd92cbbaa5f77486f3a3a06f02b42903fb5f16409998b11a WatchSource:0}: Error finding container 7e4c46378e2b20c9fd92cbbaa5f77486f3a3a06f02b42903fb5f16409998b11a: Status 404 returned error can't find the container with id 7e4c46378e2b20c9fd92cbbaa5f77486f3a3a06f02b42903fb5f16409998b11a Oct 04 10:51:44 crc kubenswrapper[5025]: I1004 10:51:44.850166 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-1b98-account-create-c872x"] Oct 04 10:51:44 crc kubenswrapper[5025]: W1004 10:51:44.857845 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod298276d3_1df7_4934_88b3_cfcbfbc0fd51.slice/crio-9b072b8ad07c247ccfec550a63aad43c29c58f7f58633e97b1689a491b65e9d3 WatchSource:0}: Error finding container 9b072b8ad07c247ccfec550a63aad43c29c58f7f58633e97b1689a491b65e9d3: Status 404 returned error can't find the container with id 9b072b8ad07c247ccfec550a63aad43c29c58f7f58633e97b1689a491b65e9d3 Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.054344 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c114-account-create-ttrlz"] Oct 04 10:51:45 crc kubenswrapper[5025]: W1004 10:51:45.060434 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce76a45b_1519_4a6e_9d7a_7a33bebadc13.slice/crio-634b0226341fcbe96b36bd70e4a6448d4cf3ed52ee8a7680140376820d24b08a WatchSource:0}: Error finding container 634b0226341fcbe96b36bd70e4a6448d4cf3ed52ee8a7680140376820d24b08a: Status 404 returned error can't find the container with id 634b0226341fcbe96b36bd70e4a6448d4cf3ed52ee8a7680140376820d24b08a Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.355807 5025 generic.go:334] "Generic (PLEG): container finished" podID="4d3aea52-35ad-49e4-93fb-465cd111f845" containerID="22c2e8b34fd6e772d90658dccb31c3f9b4484035315f76d6b01e48d3d56d0157" exitCode=0 Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.355884 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-frhrp" event={"ID":"4d3aea52-35ad-49e4-93fb-465cd111f845","Type":"ContainerDied","Data":"22c2e8b34fd6e772d90658dccb31c3f9b4484035315f76d6b01e48d3d56d0157"} Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.359887 5025 generic.go:334] "Generic (PLEG): container finished" podID="298276d3-1df7-4934-88b3-cfcbfbc0fd51" containerID="5df93cf41319e8d922adf4b51dfb0ad272c02b04679fff4cb8b2158749bf220e" exitCode=0 Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.360045 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1b98-account-create-c872x" event={"ID":"298276d3-1df7-4934-88b3-cfcbfbc0fd51","Type":"ContainerDied","Data":"5df93cf41319e8d922adf4b51dfb0ad272c02b04679fff4cb8b2158749bf220e"} Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.360074 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1b98-account-create-c872x" event={"ID":"298276d3-1df7-4934-88b3-cfcbfbc0fd51","Type":"ContainerStarted","Data":"9b072b8ad07c247ccfec550a63aad43c29c58f7f58633e97b1689a491b65e9d3"} Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.361624 5025 generic.go:334] "Generic (PLEG): container finished" podID="4323bbf4-8801-411a-8fce-bd4a077180da" containerID="5f0c2f743e61fbddf48cd901dcc0201092cbfbb21add1655d84f84894f70a55e" exitCode=0 Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.361678 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c3e5-account-create-sxqrv" event={"ID":"4323bbf4-8801-411a-8fce-bd4a077180da","Type":"ContainerDied","Data":"5f0c2f743e61fbddf48cd901dcc0201092cbfbb21add1655d84f84894f70a55e"} Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.361698 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c3e5-account-create-sxqrv" event={"ID":"4323bbf4-8801-411a-8fce-bd4a077180da","Type":"ContainerStarted","Data":"7e4c46378e2b20c9fd92cbbaa5f77486f3a3a06f02b42903fb5f16409998b11a"} Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.362857 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c114-account-create-ttrlz" event={"ID":"ce76a45b-1519-4a6e-9d7a-7a33bebadc13","Type":"ContainerStarted","Data":"c827830940a0fdc33ac2dd558980c09f9def779177524946034990802b8b36a5"} Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.362908 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c114-account-create-ttrlz" event={"ID":"ce76a45b-1519-4a6e-9d7a-7a33bebadc13","Type":"ContainerStarted","Data":"634b0226341fcbe96b36bd70e4a6448d4cf3ed52ee8a7680140376820d24b08a"} Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.365563 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="1fa2c044718960ddbde296231834fb7f5a1519308e0e7742cc723c16081ff583" exitCode=0 Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.365592 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"1fa2c044718960ddbde296231834fb7f5a1519308e0e7742cc723c16081ff583"} Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.365630 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"b41afa99f31354e6a8480ad5274a7bc72b6f827a80b5595179f0a4d79cdf33b2"} Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.365650 5025 scope.go:117] "RemoveContainer" containerID="4ca2d393da7fa0a921a67c473465c531835d20e5d5778a21a26fa2e7ee5204ae" Oct 04 10:51:45 crc kubenswrapper[5025]: I1004 10:51:45.413864 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c114-account-create-ttrlz" podStartSLOduration=1.4138419770000001 podStartE2EDuration="1.413841977s" podCreationTimestamp="2025-10-04 10:51:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:51:45.412963381 +0000 UTC m=+1033.837930281" watchObservedRunningTime="2025-10-04 10:51:45.413841977 +0000 UTC m=+1033.838808877" Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.383424 5025 generic.go:334] "Generic (PLEG): container finished" podID="ce76a45b-1519-4a6e-9d7a-7a33bebadc13" containerID="c827830940a0fdc33ac2dd558980c09f9def779177524946034990802b8b36a5" exitCode=0 Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.383864 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c114-account-create-ttrlz" event={"ID":"ce76a45b-1519-4a6e-9d7a-7a33bebadc13","Type":"ContainerDied","Data":"c827830940a0fdc33ac2dd558980c09f9def779177524946034990802b8b36a5"} Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.395688 5025 generic.go:334] "Generic (PLEG): container finished" podID="1270a45a-df30-4138-8270-78174433f9cb" containerID="149b5de7d7e8cea0b473cf87d614843df1451b44ccf2ee3c07c16f2c2af9005e" exitCode=0 Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.395752 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ctpv7" event={"ID":"1270a45a-df30-4138-8270-78174433f9cb","Type":"ContainerDied","Data":"149b5de7d7e8cea0b473cf87d614843df1451b44ccf2ee3c07c16f2c2af9005e"} Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.824727 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c3e5-account-create-sxqrv" Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.831214 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1b98-account-create-c872x" Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.844213 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.929278 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-db-sync-config-data\") pod \"4d3aea52-35ad-49e4-93fb-465cd111f845\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.929380 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-combined-ca-bundle\") pod \"4d3aea52-35ad-49e4-93fb-465cd111f845\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.929471 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2b2qx\" (UniqueName: \"kubernetes.io/projected/4323bbf4-8801-411a-8fce-bd4a077180da-kube-api-access-2b2qx\") pod \"4323bbf4-8801-411a-8fce-bd4a077180da\" (UID: \"4323bbf4-8801-411a-8fce-bd4a077180da\") " Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.929498 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll55n\" (UniqueName: \"kubernetes.io/projected/298276d3-1df7-4934-88b3-cfcbfbc0fd51-kube-api-access-ll55n\") pod \"298276d3-1df7-4934-88b3-cfcbfbc0fd51\" (UID: \"298276d3-1df7-4934-88b3-cfcbfbc0fd51\") " Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.929575 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl6rl\" (UniqueName: \"kubernetes.io/projected/4d3aea52-35ad-49e4-93fb-465cd111f845-kube-api-access-cl6rl\") pod \"4d3aea52-35ad-49e4-93fb-465cd111f845\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.929608 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-config-data\") pod \"4d3aea52-35ad-49e4-93fb-465cd111f845\" (UID: \"4d3aea52-35ad-49e4-93fb-465cd111f845\") " Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.936431 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4323bbf4-8801-411a-8fce-bd4a077180da-kube-api-access-2b2qx" (OuterVolumeSpecName: "kube-api-access-2b2qx") pod "4323bbf4-8801-411a-8fce-bd4a077180da" (UID: "4323bbf4-8801-411a-8fce-bd4a077180da"). InnerVolumeSpecName "kube-api-access-2b2qx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.936842 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d3aea52-35ad-49e4-93fb-465cd111f845-kube-api-access-cl6rl" (OuterVolumeSpecName: "kube-api-access-cl6rl") pod "4d3aea52-35ad-49e4-93fb-465cd111f845" (UID: "4d3aea52-35ad-49e4-93fb-465cd111f845"). InnerVolumeSpecName "kube-api-access-cl6rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.941262 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/298276d3-1df7-4934-88b3-cfcbfbc0fd51-kube-api-access-ll55n" (OuterVolumeSpecName: "kube-api-access-ll55n") pod "298276d3-1df7-4934-88b3-cfcbfbc0fd51" (UID: "298276d3-1df7-4934-88b3-cfcbfbc0fd51"). InnerVolumeSpecName "kube-api-access-ll55n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.941361 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4d3aea52-35ad-49e4-93fb-465cd111f845" (UID: "4d3aea52-35ad-49e4-93fb-465cd111f845"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.953492 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d3aea52-35ad-49e4-93fb-465cd111f845" (UID: "4d3aea52-35ad-49e4-93fb-465cd111f845"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:46 crc kubenswrapper[5025]: I1004 10:51:46.989917 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-config-data" (OuterVolumeSpecName: "config-data") pod "4d3aea52-35ad-49e4-93fb-465cd111f845" (UID: "4d3aea52-35ad-49e4-93fb-465cd111f845"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.031797 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl6rl\" (UniqueName: \"kubernetes.io/projected/4d3aea52-35ad-49e4-93fb-465cd111f845-kube-api-access-cl6rl\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.031823 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.031832 5025 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.031841 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d3aea52-35ad-49e4-93fb-465cd111f845-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.031850 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2b2qx\" (UniqueName: \"kubernetes.io/projected/4323bbf4-8801-411a-8fce-bd4a077180da-kube-api-access-2b2qx\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.031858 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll55n\" (UniqueName: \"kubernetes.io/projected/298276d3-1df7-4934-88b3-cfcbfbc0fd51-kube-api-access-ll55n\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.433169 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-frhrp" event={"ID":"4d3aea52-35ad-49e4-93fb-465cd111f845","Type":"ContainerDied","Data":"ab7d8bd00691ce068eb7c7ab6508b8d56a009a3cccad3ceafbb2e51ac8885c80"} Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.433258 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab7d8bd00691ce068eb7c7ab6508b8d56a009a3cccad3ceafbb2e51ac8885c80" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.433465 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-frhrp" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.441764 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-1b98-account-create-c872x" event={"ID":"298276d3-1df7-4934-88b3-cfcbfbc0fd51","Type":"ContainerDied","Data":"9b072b8ad07c247ccfec550a63aad43c29c58f7f58633e97b1689a491b65e9d3"} Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.441830 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b072b8ad07c247ccfec550a63aad43c29c58f7f58633e97b1689a491b65e9d3" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.441967 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-1b98-account-create-c872x" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.444182 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c3e5-account-create-sxqrv" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.444268 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c3e5-account-create-sxqrv" event={"ID":"4323bbf4-8801-411a-8fce-bd4a077180da","Type":"ContainerDied","Data":"7e4c46378e2b20c9fd92cbbaa5f77486f3a3a06f02b42903fb5f16409998b11a"} Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.444297 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e4c46378e2b20c9fd92cbbaa5f77486f3a3a06f02b42903fb5f16409998b11a" Oct 04 10:51:47 crc kubenswrapper[5025]: E1004 10:51:47.601706 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d3aea52_35ad_49e4_93fb_465cd111f845.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod298276d3_1df7_4934_88b3_cfcbfbc0fd51.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d3aea52_35ad_49e4_93fb_465cd111f845.slice/crio-ab7d8bd00691ce068eb7c7ab6508b8d56a009a3cccad3ceafbb2e51ac8885c80\": RecentStats: unable to find data in memory cache]" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.819818 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-h44cd"] Oct 04 10:51:47 crc kubenswrapper[5025]: E1004 10:51:47.821412 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d3aea52-35ad-49e4-93fb-465cd111f845" containerName="glance-db-sync" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.821431 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d3aea52-35ad-49e4-93fb-465cd111f845" containerName="glance-db-sync" Oct 04 10:51:47 crc kubenswrapper[5025]: E1004 10:51:47.821464 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4323bbf4-8801-411a-8fce-bd4a077180da" containerName="mariadb-account-create" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.821469 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="4323bbf4-8801-411a-8fce-bd4a077180da" containerName="mariadb-account-create" Oct 04 10:51:47 crc kubenswrapper[5025]: E1004 10:51:47.821486 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="298276d3-1df7-4934-88b3-cfcbfbc0fd51" containerName="mariadb-account-create" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.821492 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="298276d3-1df7-4934-88b3-cfcbfbc0fd51" containerName="mariadb-account-create" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.821646 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d3aea52-35ad-49e4-93fb-465cd111f845" containerName="glance-db-sync" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.821666 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="298276d3-1df7-4934-88b3-cfcbfbc0fd51" containerName="mariadb-account-create" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.821679 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="4323bbf4-8801-411a-8fce-bd4a077180da" containerName="mariadb-account-create" Oct 04 10:51:47 crc kubenswrapper[5025]: I1004 10:51:47.829177 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:47.872783 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-h44cd"] Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:47.913814 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:47.951891 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-svc\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:47.951939 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:47.952039 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4wlz\" (UniqueName: \"kubernetes.io/projected/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-kube-api-access-f4wlz\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:47.952070 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:47.952104 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-config\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:47.952250 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:47.990653 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c114-account-create-ttrlz" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.053703 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8vsc\" (UniqueName: \"kubernetes.io/projected/1270a45a-df30-4138-8270-78174433f9cb-kube-api-access-v8vsc\") pod \"1270a45a-df30-4138-8270-78174433f9cb\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.053812 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-combined-ca-bundle\") pod \"1270a45a-df30-4138-8270-78174433f9cb\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.053928 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-config-data\") pod \"1270a45a-df30-4138-8270-78174433f9cb\" (UID: \"1270a45a-df30-4138-8270-78174433f9cb\") " Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.061250 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4wlz\" (UniqueName: \"kubernetes.io/projected/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-kube-api-access-f4wlz\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.061279 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.061311 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-config\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.061419 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.061480 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-svc\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.061507 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.062381 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-nb\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.062965 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-config\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.063727 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-swift-storage-0\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.064212 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-sb\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.064264 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-svc\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.069510 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1270a45a-df30-4138-8270-78174433f9cb-kube-api-access-v8vsc" (OuterVolumeSpecName: "kube-api-access-v8vsc") pod "1270a45a-df30-4138-8270-78174433f9cb" (UID: "1270a45a-df30-4138-8270-78174433f9cb"). InnerVolumeSpecName "kube-api-access-v8vsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.084289 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1270a45a-df30-4138-8270-78174433f9cb" (UID: "1270a45a-df30-4138-8270-78174433f9cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.085545 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4wlz\" (UniqueName: \"kubernetes.io/projected/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-kube-api-access-f4wlz\") pod \"dnsmasq-dns-895cf5cf-h44cd\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.119167 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-config-data" (OuterVolumeSpecName: "config-data") pod "1270a45a-df30-4138-8270-78174433f9cb" (UID: "1270a45a-df30-4138-8270-78174433f9cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.162559 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g74d9\" (UniqueName: \"kubernetes.io/projected/ce76a45b-1519-4a6e-9d7a-7a33bebadc13-kube-api-access-g74d9\") pod \"ce76a45b-1519-4a6e-9d7a-7a33bebadc13\" (UID: \"ce76a45b-1519-4a6e-9d7a-7a33bebadc13\") " Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.163201 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.163218 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1270a45a-df30-4138-8270-78174433f9cb-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.163227 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8vsc\" (UniqueName: \"kubernetes.io/projected/1270a45a-df30-4138-8270-78174433f9cb-kube-api-access-v8vsc\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.166186 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce76a45b-1519-4a6e-9d7a-7a33bebadc13-kube-api-access-g74d9" (OuterVolumeSpecName: "kube-api-access-g74d9") pod "ce76a45b-1519-4a6e-9d7a-7a33bebadc13" (UID: "ce76a45b-1519-4a6e-9d7a-7a33bebadc13"). InnerVolumeSpecName "kube-api-access-g74d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.215610 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.264812 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g74d9\" (UniqueName: \"kubernetes.io/projected/ce76a45b-1519-4a6e-9d7a-7a33bebadc13-kube-api-access-g74d9\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.454543 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ctpv7" event={"ID":"1270a45a-df30-4138-8270-78174433f9cb","Type":"ContainerDied","Data":"bac07a0bfb19984d0c026ce7af78618628555a93e776211eb7b8955c571475a5"} Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.454579 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bac07a0bfb19984d0c026ce7af78618628555a93e776211eb7b8955c571475a5" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.454645 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ctpv7" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.456314 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c114-account-create-ttrlz" event={"ID":"ce76a45b-1519-4a6e-9d7a-7a33bebadc13","Type":"ContainerDied","Data":"634b0226341fcbe96b36bd70e4a6448d4cf3ed52ee8a7680140376820d24b08a"} Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.456343 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="634b0226341fcbe96b36bd70e4a6448d4cf3ed52ee8a7680140376820d24b08a" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.456330 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c114-account-create-ttrlz" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.676348 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-h44cd"] Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.741160 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hsvbc"] Oct 04 10:51:48 crc kubenswrapper[5025]: E1004 10:51:48.741548 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1270a45a-df30-4138-8270-78174433f9cb" containerName="keystone-db-sync" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.741572 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="1270a45a-df30-4138-8270-78174433f9cb" containerName="keystone-db-sync" Oct 04 10:51:48 crc kubenswrapper[5025]: E1004 10:51:48.741604 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce76a45b-1519-4a6e-9d7a-7a33bebadc13" containerName="mariadb-account-create" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.741614 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce76a45b-1519-4a6e-9d7a-7a33bebadc13" containerName="mariadb-account-create" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.741809 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce76a45b-1519-4a6e-9d7a-7a33bebadc13" containerName="mariadb-account-create" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.741841 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="1270a45a-df30-4138-8270-78174433f9cb" containerName="keystone-db-sync" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.742548 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.748039 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.748353 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.748541 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.748692 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9d9vx" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.754071 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hsvbc"] Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.775292 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-jmmdd"] Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.776636 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.814597 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-jmmdd"] Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875063 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkh7d\" (UniqueName: \"kubernetes.io/projected/302a0707-59c7-46af-a5dc-cc3c6fa23889-kube-api-access-tkh7d\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875112 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-config-data\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875139 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875155 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-config\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875193 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875213 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-fernet-keys\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875258 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzz7h\" (UniqueName: \"kubernetes.io/projected/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-kube-api-access-kzz7h\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875274 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-scripts\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875291 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-credential-keys\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875312 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875333 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.875369 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-combined-ca-bundle\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.929029 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-68bbfdcf9-mllx9"] Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.939902 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.944253 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.944338 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-lwxwq" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.944432 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.947744 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.991504 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzz7h\" (UniqueName: \"kubernetes.io/projected/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-kube-api-access-kzz7h\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.991576 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-scripts\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.991610 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-credential-keys\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.991671 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.991715 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.991824 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-combined-ca-bundle\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.991878 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkh7d\" (UniqueName: \"kubernetes.io/projected/302a0707-59c7-46af-a5dc-cc3c6fa23889-kube-api-access-tkh7d\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.991925 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-config-data\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.991964 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-config\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.991988 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.992082 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:48 crc kubenswrapper[5025]: I1004 10:51:48.992126 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-fernet-keys\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:48.999232 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-svc\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:48.999800 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-nb\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.011682 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-config\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.013229 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-combined-ca-bundle\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.013296 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68bbfdcf9-mllx9"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.015485 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-swift-storage-0\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.019036 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-sb\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.022873 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-config-data\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.041567 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-credential-keys\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.049742 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzz7h\" (UniqueName: \"kubernetes.io/projected/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-kube-api-access-kzz7h\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.050694 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-scripts\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.055197 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-fernet-keys\") pod \"keystone-bootstrap-hsvbc\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.067590 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.077977 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkh7d\" (UniqueName: \"kubernetes.io/projected/302a0707-59c7-46af-a5dc-cc3c6fa23889-kube-api-access-tkh7d\") pod \"dnsmasq-dns-6c9c9f998c-jmmdd\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.102467 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.103623 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2b5229-8605-401c-9321-58b64cd1262a-logs\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.103678 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-scripts\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.103734 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnf6m\" (UniqueName: \"kubernetes.io/projected/9d2b5229-8605-401c-9321-58b64cd1262a-kube-api-access-vnf6m\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.103763 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d2b5229-8605-401c-9321-58b64cd1262a-horizon-secret-key\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.103793 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-config-data\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.108140 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-h44cd"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.194586 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-g94ln"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.195912 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.214246 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d2b5229-8605-401c-9321-58b64cd1262a-horizon-secret-key\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.214294 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-config-data\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.214332 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/565cc52d-cdcf-4d59-82b5-3d71804b727a-logs\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.214352 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xbfz\" (UniqueName: \"kubernetes.io/projected/565cc52d-cdcf-4d59-82b5-3d71804b727a-kube-api-access-7xbfz\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.214403 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2b5229-8605-401c-9321-58b64cd1262a-logs\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.214424 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-scripts\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.214447 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-scripts\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.214473 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-config-data\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.214493 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-combined-ca-bundle\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.214509 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnf6m\" (UniqueName: \"kubernetes.io/projected/9d2b5229-8605-401c-9321-58b64cd1262a-kube-api-access-vnf6m\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.215434 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-mc766" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.215564 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.215672 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.216925 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2b5229-8605-401c-9321-58b64cd1262a-logs\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.222486 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d2b5229-8605-401c-9321-58b64cd1262a-horizon-secret-key\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.222789 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-scripts\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.234115 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-g94ln"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.234911 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-config-data\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.243360 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnf6m\" (UniqueName: \"kubernetes.io/projected/9d2b5229-8605-401c-9321-58b64cd1262a-kube-api-access-vnf6m\") pod \"horizon-68bbfdcf9-mllx9\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.247816 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-jmmdd"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.279005 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.288224 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-tspdr"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.297930 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.317920 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-config-data\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.317983 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-combined-ca-bundle\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.318075 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/565cc52d-cdcf-4d59-82b5-3d71804b727a-logs\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.318108 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xbfz\" (UniqueName: \"kubernetes.io/projected/565cc52d-cdcf-4d59-82b5-3d71804b727a-kube-api-access-7xbfz\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.318186 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-scripts\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.320990 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/565cc52d-cdcf-4d59-82b5-3d71804b727a-logs\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.322149 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-scripts\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.325496 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-config-data\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.328593 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-combined-ca-bundle\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.336070 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-tspdr"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.359039 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.361406 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.364673 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xbfz\" (UniqueName: \"kubernetes.io/projected/565cc52d-cdcf-4d59-82b5-3d71804b727a-kube-api-access-7xbfz\") pod \"placement-db-sync-g94ln\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.385473 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.387905 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.422271 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwtpd\" (UniqueName: \"kubernetes.io/projected/510866f0-02ee-48ef-8219-fdea80015c5a-kube-api-access-mwtpd\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.422334 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.422396 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.422419 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.422455 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.422509 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-config\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.429663 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.431073 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.439131 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4867c" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.439325 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.439487 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.476082 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.484184 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7cbc48db97-z46n4"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.485880 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.491775 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cbc48db97-z46n4"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.493742 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-h44cd" event={"ID":"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703","Type":"ContainerStarted","Data":"9fe3b0132bd909954a3df494885baec2592a6cf250525158975056d2e164fe12"} Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.511359 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-sfm5z"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.514384 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.524543 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.524803 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-27v2q" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.525807 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbbdh\" (UniqueName: \"kubernetes.io/projected/03dee37f-5b63-4b21-b401-f05b07d3b1af-kube-api-access-wbbdh\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.525866 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-log-httpd\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.525898 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.525940 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwtpd\" (UniqueName: \"kubernetes.io/projected/510866f0-02ee-48ef-8219-fdea80015c5a-kube-api-access-mwtpd\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.526005 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.526070 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-scripts\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.526110 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.526178 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-config-data\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.526205 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.526229 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.526268 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.526334 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-config\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.526379 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-run-httpd\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.534651 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.535344 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.535745 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.538069 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.542290 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.544054 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-config\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.567520 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sfm5z"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.569907 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g94ln" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.585071 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwtpd\" (UniqueName: \"kubernetes.io/projected/510866f0-02ee-48ef-8219-fdea80015c5a-kube-api-access-mwtpd\") pod \"dnsmasq-dns-57c957c4ff-tspdr\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.617769 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-xtspb"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.618947 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.622695 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jnpdb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.623455 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.623750 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.634756 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.634881 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-config-data\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.634912 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k86p\" (UniqueName: \"kubernetes.io/projected/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-kube-api-access-7k86p\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.634946 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b4f50f0c-193b-4e77-a778-2098ba1f756f-horizon-secret-key\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.634976 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.634998 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-config-data\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635040 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635059 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-run-httpd\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635075 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbbdh\" (UniqueName: \"kubernetes.io/projected/03dee37f-5b63-4b21-b401-f05b07d3b1af-kube-api-access-wbbdh\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635099 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7mnr\" (UniqueName: \"kubernetes.io/projected/b5d127d5-e426-48fe-bb22-c811de5c801d-kube-api-access-w7mnr\") pod \"barbican-db-sync-sfm5z\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635115 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-db-sync-config-data\") pod \"barbican-db-sync-sfm5z\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635130 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-config-data\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635146 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-log-httpd\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635163 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-scripts\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635180 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635216 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kxcp\" (UniqueName: \"kubernetes.io/projected/b4f50f0c-193b-4e77-a778-2098ba1f756f-kube-api-access-4kxcp\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635329 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635360 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-logs\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635379 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-scripts\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635398 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4f50f0c-193b-4e77-a778-2098ba1f756f-logs\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635411 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-combined-ca-bundle\") pod \"barbican-db-sync-sfm5z\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.635431 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-scripts\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.636706 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-log-httpd\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.638916 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-run-httpd\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.639876 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-scripts\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.641750 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-config-data\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.647473 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.652568 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.661890 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbbdh\" (UniqueName: \"kubernetes.io/projected/03dee37f-5b63-4b21-b401-f05b07d3b1af-kube-api-access-wbbdh\") pod \"ceilometer-0\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.665755 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xtspb"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.694318 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.726164 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-25rkf"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.727885 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.732345 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.738874 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b4f50f0c-193b-4e77-a778-2098ba1f756f-horizon-secret-key\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.738932 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.738940 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.738953 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68hzg\" (UniqueName: \"kubernetes.io/projected/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-kube-api-access-68hzg\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.738979 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-config-data\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739000 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-db-sync-config-data\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739037 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739061 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-combined-ca-bundle\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739086 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7mnr\" (UniqueName: \"kubernetes.io/projected/b5d127d5-e426-48fe-bb22-c811de5c801d-kube-api-access-w7mnr\") pod \"barbican-db-sync-sfm5z\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739101 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-db-sync-config-data\") pod \"barbican-db-sync-sfm5z\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739119 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-config-data\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739139 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-scripts\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739167 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-config-data\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739185 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-etc-machine-id\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739250 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739272 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kxcp\" (UniqueName: \"kubernetes.io/projected/b4f50f0c-193b-4e77-a778-2098ba1f756f-kube-api-access-4kxcp\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739298 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-logs\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739319 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4f50f0c-193b-4e77-a778-2098ba1f756f-logs\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739334 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-combined-ca-bundle\") pod \"barbican-db-sync-sfm5z\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739352 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-scripts\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739384 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-scripts\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.739400 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k86p\" (UniqueName: \"kubernetes.io/projected/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-kube-api-access-7k86p\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.740326 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-d8fx9" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.741058 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.742638 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-scripts\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.742929 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b4f50f0c-193b-4e77-a778-2098ba1f756f-horizon-secret-key\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.744677 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.746867 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.756477 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-logs\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.756709 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4f50f0c-193b-4e77-a778-2098ba1f756f-logs\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.762077 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-scripts\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.762693 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.763227 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-combined-ca-bundle\") pod \"barbican-db-sync-sfm5z\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.763789 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k86p\" (UniqueName: \"kubernetes.io/projected/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-kube-api-access-7k86p\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.764413 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-config-data\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.781123 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-25rkf"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.786962 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-config-data\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.787367 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-db-sync-config-data\") pod \"barbican-db-sync-sfm5z\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.787862 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kxcp\" (UniqueName: \"kubernetes.io/projected/b4f50f0c-193b-4e77-a778-2098ba1f756f-kube-api-access-4kxcp\") pod \"horizon-7cbc48db97-z46n4\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.798069 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " pod="openstack/glance-default-external-api-0" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.814456 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7mnr\" (UniqueName: \"kubernetes.io/projected/b5d127d5-e426-48fe-bb22-c811de5c801d-kube-api-access-w7mnr\") pod \"barbican-db-sync-sfm5z\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.828582 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.843075 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-db-sync-config-data\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.843128 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-combined-ca-bundle\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.843175 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-combined-ca-bundle\") pod \"neutron-db-sync-25rkf\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.843202 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-config-data\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.843222 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-etc-machine-id\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.843239 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-config\") pod \"neutron-db-sync-25rkf\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.843274 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rs59\" (UniqueName: \"kubernetes.io/projected/8e307d63-f22a-4e97-9385-773e9ce7f4fd-kube-api-access-4rs59\") pod \"neutron-db-sync-25rkf\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.843313 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-scripts\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.843353 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68hzg\" (UniqueName: \"kubernetes.io/projected/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-kube-api-access-68hzg\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.845106 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-etc-machine-id\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.857495 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.871897 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-scripts\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.872421 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-config-data\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.886326 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-db-sync-config-data\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.912035 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68hzg\" (UniqueName: \"kubernetes.io/projected/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-kube-api-access-68hzg\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.914059 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hsvbc"] Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.915155 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-combined-ca-bundle\") pod \"cinder-db-sync-xtspb\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.924664 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xtspb" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.945149 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-combined-ca-bundle\") pod \"neutron-db-sync-25rkf\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.945501 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-config\") pod \"neutron-db-sync-25rkf\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.945548 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rs59\" (UniqueName: \"kubernetes.io/projected/8e307d63-f22a-4e97-9385-773e9ce7f4fd-kube-api-access-4rs59\") pod \"neutron-db-sync-25rkf\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.949785 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-combined-ca-bundle\") pod \"neutron-db-sync-25rkf\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.949971 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-config\") pod \"neutron-db-sync-25rkf\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:49 crc kubenswrapper[5025]: I1004 10:51:49.961740 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rs59\" (UniqueName: \"kubernetes.io/projected/8e307d63-f22a-4e97-9385-773e9ce7f4fd-kube-api-access-4rs59\") pod \"neutron-db-sync-25rkf\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.063051 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-jmmdd"] Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.068206 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.165911 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-g94ln"] Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.200607 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.202200 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.204432 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.209527 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.219620 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-68bbfdcf9-mllx9"] Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.240162 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-25rkf" Oct 04 10:51:50 crc kubenswrapper[5025]: W1004 10:51:50.252299 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d2b5229_8605_401c_9321_58b64cd1262a.slice/crio-9fe90c4608846605874c26c6faa0bfdc97b95c511d6c3d10ea36d872ceed2bf7 WatchSource:0}: Error finding container 9fe90c4608846605874c26c6faa0bfdc97b95c511d6c3d10ea36d872ceed2bf7: Status 404 returned error can't find the container with id 9fe90c4608846605874c26c6faa0bfdc97b95c511d6c3d10ea36d872ceed2bf7 Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.355967 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.356303 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-logs\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.356328 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.356365 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph9rb\" (UniqueName: \"kubernetes.io/projected/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-kube-api-access-ph9rb\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.356393 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.356420 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.356442 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.457599 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.457647 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.457669 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.457750 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.457772 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-logs\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.457797 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.457826 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph9rb\" (UniqueName: \"kubernetes.io/projected/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-kube-api-access-ph9rb\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.458534 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.459504 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-logs\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.459537 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.468365 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.468471 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.469929 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.479802 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph9rb\" (UniqueName: \"kubernetes.io/projected/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-kube-api-access-ph9rb\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.502480 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.511337 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-tspdr"] Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.518369 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7cbc48db97-z46n4"] Oct 04 10:51:50 crc kubenswrapper[5025]: W1004 10:51:50.519370 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4f50f0c_193b_4e77_a778_2098ba1f756f.slice/crio-5a8772e3f6c17cbcb214c9a46fe8b04870d133a173aea59e108275bc40ddf22b WatchSource:0}: Error finding container 5a8772e3f6c17cbcb214c9a46fe8b04870d133a173aea59e108275bc40ddf22b: Status 404 returned error can't find the container with id 5a8772e3f6c17cbcb214c9a46fe8b04870d133a173aea59e108275bc40ddf22b Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.523049 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: W1004 10:51:50.524908 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod510866f0_02ee_48ef_8219_fdea80015c5a.slice/crio-05112581d44898b9a0f956529afc46d25f85b2c6f4f62a85b4c443073d7fb567 WatchSource:0}: Error finding container 05112581d44898b9a0f956529afc46d25f85b2c6f4f62a85b4c443073d7fb567: Status 404 returned error can't find the container with id 05112581d44898b9a0f956529afc46d25f85b2c6f4f62a85b4c443073d7fb567 Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.531613 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsvbc" event={"ID":"5b4eee55-362c-41ba-bc6d-4b29c77bddd5","Type":"ContainerStarted","Data":"d6cafb614f4d4c7ccb64f8e00d276df76d61213552f09efaff5591b1e84d47d1"} Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.531653 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsvbc" event={"ID":"5b4eee55-362c-41ba-bc6d-4b29c77bddd5","Type":"ContainerStarted","Data":"2a7b9c5ae51fa0987c4f55ae4fa95e488699a014eade5e2c8412d7fff1734b1a"} Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.540408 5025 generic.go:334] "Generic (PLEG): container finished" podID="e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" containerID="ef27343331976d22ee7c4365ecece05d2ce0fda141f21185749ed8e2eafed52b" exitCode=0 Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.540617 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-h44cd" event={"ID":"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703","Type":"ContainerDied","Data":"ef27343331976d22ee7c4365ecece05d2ce0fda141f21185749ed8e2eafed52b"} Oct 04 10:51:50 crc kubenswrapper[5025]: W1004 10:51:50.548374 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5d127d5_e426_48fe_bb22_c811de5c801d.slice/crio-6ece5abf1e402fe3af16910eaee087f8451761283ab612edac0c69c5c6907eda WatchSource:0}: Error finding container 6ece5abf1e402fe3af16910eaee087f8451761283ab612edac0c69c5c6907eda: Status 404 returned error can't find the container with id 6ece5abf1e402fe3af16910eaee087f8451761283ab612edac0c69c5c6907eda Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.552275 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-sfm5z"] Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.554791 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hsvbc" podStartSLOduration=2.554776199 podStartE2EDuration="2.554776199s" podCreationTimestamp="2025-10-04 10:51:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:51:50.548073833 +0000 UTC m=+1038.973040703" watchObservedRunningTime="2025-10-04 10:51:50.554776199 +0000 UTC m=+1038.979743079" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.557304 5025 generic.go:334] "Generic (PLEG): container finished" podID="302a0707-59c7-46af-a5dc-cc3c6fa23889" containerID="abc8cd96d9506bcf2a7ee640a2581500e305787452bca3102bfff8d674ff8508" exitCode=0 Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.557519 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" event={"ID":"302a0707-59c7-46af-a5dc-cc3c6fa23889","Type":"ContainerDied","Data":"abc8cd96d9506bcf2a7ee640a2581500e305787452bca3102bfff8d674ff8508"} Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.557554 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" event={"ID":"302a0707-59c7-46af-a5dc-cc3c6fa23889","Type":"ContainerStarted","Data":"9bdc9093c017342d5d3ce0d5473c8bdb9c77ef545aca5070cc724e1f0006a0c9"} Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.561951 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68bbfdcf9-mllx9" event={"ID":"9d2b5229-8605-401c-9321-58b64cd1262a","Type":"ContainerStarted","Data":"9fe90c4608846605874c26c6faa0bfdc97b95c511d6c3d10ea36d872ceed2bf7"} Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.563087 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g94ln" event={"ID":"565cc52d-cdcf-4d59-82b5-3d71804b727a","Type":"ContainerStarted","Data":"850bf24e3e048dc319f2e926d50f276578ff822b45a927590b6c20f9d74f6ced"} Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.714378 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-xtspb"] Oct 04 10:51:50 crc kubenswrapper[5025]: W1004 10:51:50.718748 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod694e381e_49b8_49f1_a5ae_7aa3b72d8fee.slice/crio-c03095558b2e2a07df1cf4fb37ac72b9357c6c9d45470610d2cf0bbf0ab11242 WatchSource:0}: Error finding container c03095558b2e2a07df1cf4fb37ac72b9357c6c9d45470610d2cf0bbf0ab11242: Status 404 returned error can't find the container with id c03095558b2e2a07df1cf4fb37ac72b9357c6c9d45470610d2cf0bbf0ab11242 Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.817929 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.826977 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-25rkf"] Oct 04 10:51:50 crc kubenswrapper[5025]: I1004 10:51:50.954476 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.102779 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.171736 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-nb\") pod \"302a0707-59c7-46af-a5dc-cc3c6fa23889\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.171792 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-sb\") pod \"302a0707-59c7-46af-a5dc-cc3c6fa23889\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.171851 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-config\") pod \"302a0707-59c7-46af-a5dc-cc3c6fa23889\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.171906 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkh7d\" (UniqueName: \"kubernetes.io/projected/302a0707-59c7-46af-a5dc-cc3c6fa23889-kube-api-access-tkh7d\") pod \"302a0707-59c7-46af-a5dc-cc3c6fa23889\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.171942 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-swift-storage-0\") pod \"302a0707-59c7-46af-a5dc-cc3c6fa23889\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.171982 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-svc\") pod \"302a0707-59c7-46af-a5dc-cc3c6fa23889\" (UID: \"302a0707-59c7-46af-a5dc-cc3c6fa23889\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.189759 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.193886 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/302a0707-59c7-46af-a5dc-cc3c6fa23889-kube-api-access-tkh7d" (OuterVolumeSpecName: "kube-api-access-tkh7d") pod "302a0707-59c7-46af-a5dc-cc3c6fa23889" (UID: "302a0707-59c7-46af-a5dc-cc3c6fa23889"). InnerVolumeSpecName "kube-api-access-tkh7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.221536 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "302a0707-59c7-46af-a5dc-cc3c6fa23889" (UID: "302a0707-59c7-46af-a5dc-cc3c6fa23889"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.225441 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-config" (OuterVolumeSpecName: "config") pod "302a0707-59c7-46af-a5dc-cc3c6fa23889" (UID: "302a0707-59c7-46af-a5dc-cc3c6fa23889"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.228593 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "302a0707-59c7-46af-a5dc-cc3c6fa23889" (UID: "302a0707-59c7-46af-a5dc-cc3c6fa23889"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.246046 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "302a0707-59c7-46af-a5dc-cc3c6fa23889" (UID: "302a0707-59c7-46af-a5dc-cc3c6fa23889"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.251664 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "302a0707-59c7-46af-a5dc-cc3c6fa23889" (UID: "302a0707-59c7-46af-a5dc-cc3c6fa23889"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.273619 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-swift-storage-0\") pod \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.273662 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-nb\") pod \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.273744 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-sb\") pod \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.273816 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4wlz\" (UniqueName: \"kubernetes.io/projected/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-kube-api-access-f4wlz\") pod \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.273874 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-svc\") pod \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.273894 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-config\") pod \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\" (UID: \"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703\") " Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.274233 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.274245 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkh7d\" (UniqueName: \"kubernetes.io/projected/302a0707-59c7-46af-a5dc-cc3c6fa23889-kube-api-access-tkh7d\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.274254 5025 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.274271 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.274280 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.274288 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/302a0707-59c7-46af-a5dc-cc3c6fa23889-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.286934 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-kube-api-access-f4wlz" (OuterVolumeSpecName: "kube-api-access-f4wlz") pod "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" (UID: "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703"). InnerVolumeSpecName "kube-api-access-f4wlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.298486 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" (UID: "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.302555 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" (UID: "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: W1004 10:51:51.303744 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a11a8b2_3ed8_4978_b7ed_298dba8db93d.slice/crio-843a03026ef2521cbfd72cf87dc1e0472fd7f53fed109a3080292f49e4fd6c8d WatchSource:0}: Error finding container 843a03026ef2521cbfd72cf87dc1e0472fd7f53fed109a3080292f49e4fd6c8d: Status 404 returned error can't find the container with id 843a03026ef2521cbfd72cf87dc1e0472fd7f53fed109a3080292f49e4fd6c8d Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.308381 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.316534 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" (UID: "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: E1004 10:51:51.334910 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = reading blob sha256:00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/00/00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251004%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251004T105151Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=2cff85343eb4d8be99e2501f1cb0ba779e0019782358b2fecbd407c6e59d2da9®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-horizon&akamai_signature=exp=1759576011~hmac=4650935ff1f610e28792266a8d1566f0c0faab1b337959d3951f589a9d016394\": EOF" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 10:51:51 crc kubenswrapper[5025]: E1004 10:51:51.335804 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc5h5b9h8dh5c5h56dh548h8ch558h686h66dh5c7h548h59h8bhc8h5dfhd6h5f7h567h687h65dh5fdh695h585h694h5d7h85h5f9h546h8bhc9h5b4q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4kxcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7cbc48db97-z46n4_openstack(b4f50f0c-193b-4e77-a778-2098ba1f756f): ErrImagePull: reading blob sha256:00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f: Get \"https://cdn01.quay.io/quayio-production-s3/sha256/00/00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251004%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251004T105151Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=2cff85343eb4d8be99e2501f1cb0ba779e0019782358b2fecbd407c6e59d2da9®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-horizon&akamai_signature=exp=1759576011~hmac=4650935ff1f610e28792266a8d1566f0c0faab1b337959d3951f589a9d016394\": EOF" logger="UnhandledError" Oct 04 10:51:51 crc kubenswrapper[5025]: E1004 10:51:51.341787 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"reading blob sha256:00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f: Get \\\"https://cdn01.quay.io/quayio-production-s3/sha256/00/00c7dbee21a6412f55c100744d2ab999ceb7298f3450d543101adfb9d78abc6f?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIATAAF2YHTGR23ZTE6%2F20251004%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20251004T105151Z&X-Amz-Expires=600&X-Amz-SignedHeaders=host&X-Amz-Signature=2cff85343eb4d8be99e2501f1cb0ba779e0019782358b2fecbd407c6e59d2da9®ion=us-east-1&namespace=podified-antelope-centos9&username=openshift-release-dev+ocm_access_1b89217552bc42d1be3fb06a1aed001a&repo_name=openstack-horizon&akamai_signature=exp=1759576011~hmac=4650935ff1f610e28792266a8d1566f0c0faab1b337959d3951f589a9d016394\\\": EOF\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7cbc48db97-z46n4" podUID="b4f50f0c-193b-4e77-a778-2098ba1f756f" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.374447 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" (UID: "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.380762 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-config" (OuterVolumeSpecName: "config") pod "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" (UID: "e27cdce2-cc3d-4643-b0ce-a7c2f2acf703"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.383124 5025 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.383157 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.383169 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.383178 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4wlz\" (UniqueName: \"kubernetes.io/projected/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-kube-api-access-f4wlz\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.383189 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.383203 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.573282 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-895cf5cf-h44cd" event={"ID":"e27cdce2-cc3d-4643-b0ce-a7c2f2acf703","Type":"ContainerDied","Data":"9fe3b0132bd909954a3df494885baec2592a6cf250525158975056d2e164fe12"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.573330 5025 scope.go:117] "RemoveContainer" containerID="ef27343331976d22ee7c4365ecece05d2ce0fda141f21185749ed8e2eafed52b" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.573384 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-895cf5cf-h44cd" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.575640 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" event={"ID":"302a0707-59c7-46af-a5dc-cc3c6fa23889","Type":"ContainerDied","Data":"9bdc9093c017342d5d3ce0d5473c8bdb9c77ef545aca5070cc724e1f0006a0c9"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.575728 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c9c9f998c-jmmdd" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.579345 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03dee37f-5b63-4b21-b401-f05b07d3b1af","Type":"ContainerStarted","Data":"748fad10349346b8ecb24a1a99bfd2ff129149110b4144764266f0f137b83244"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.588692 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cbc48db97-z46n4" event={"ID":"b4f50f0c-193b-4e77-a778-2098ba1f756f","Type":"ContainerStarted","Data":"5a8772e3f6c17cbcb214c9a46fe8b04870d133a173aea59e108275bc40ddf22b"} Oct 04 10:51:51 crc kubenswrapper[5025]: E1004 10:51:51.591629 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7cbc48db97-z46n4" podUID="b4f50f0c-193b-4e77-a778-2098ba1f756f" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.597211 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-25rkf" event={"ID":"8e307d63-f22a-4e97-9385-773e9ce7f4fd","Type":"ContainerStarted","Data":"6bb3b61562fd3fec493b942be4e930f472a30c9bb280ebef518fa016041ed7b7"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.597256 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-25rkf" event={"ID":"8e307d63-f22a-4e97-9385-773e9ce7f4fd","Type":"ContainerStarted","Data":"0070e3e8dda26726f80f094b06496edcff93989e8c55dcf61cb0be944e4e59a3"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.605784 5025 generic.go:334] "Generic (PLEG): container finished" podID="510866f0-02ee-48ef-8219-fdea80015c5a" containerID="1c155077832c655cf8b156dff2fa8d96206cca3c55adff9fef5052513b07d4e4" exitCode=0 Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.605915 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" event={"ID":"510866f0-02ee-48ef-8219-fdea80015c5a","Type":"ContainerDied","Data":"1c155077832c655cf8b156dff2fa8d96206cca3c55adff9fef5052513b07d4e4"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.605944 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" event={"ID":"510866f0-02ee-48ef-8219-fdea80015c5a","Type":"ContainerStarted","Data":"05112581d44898b9a0f956529afc46d25f85b2c6f4f62a85b4c443073d7fb567"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.609325 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sfm5z" event={"ID":"b5d127d5-e426-48fe-bb22-c811de5c801d","Type":"ContainerStarted","Data":"6ece5abf1e402fe3af16910eaee087f8451761283ab612edac0c69c5c6907eda"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.612190 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b4531e3d-3ed8-411c-ae2e-5c0893d250b5","Type":"ContainerStarted","Data":"479a50bb495ab54164d313abef8cc81523b09af4b7949fa6a7c7aa680a2b07e8"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.613320 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a11a8b2-3ed8-4978-b7ed-298dba8db93d","Type":"ContainerStarted","Data":"843a03026ef2521cbfd72cf87dc1e0472fd7f53fed109a3080292f49e4fd6c8d"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.615813 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xtspb" event={"ID":"694e381e-49b8-49f1-a5ae-7aa3b72d8fee","Type":"ContainerStarted","Data":"c03095558b2e2a07df1cf4fb37ac72b9357c6c9d45470610d2cf0bbf0ab11242"} Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.631508 5025 scope.go:117] "RemoveContainer" containerID="abc8cd96d9506bcf2a7ee640a2581500e305787452bca3102bfff8d674ff8508" Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.645882 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-h44cd"] Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.652923 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-895cf5cf-h44cd"] Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.675825 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-jmmdd"] Oct 04 10:51:51 crc kubenswrapper[5025]: I1004 10:51:51.680867 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c9c9f998c-jmmdd"] Oct 04 10:51:52 crc kubenswrapper[5025]: I1004 10:51:52.428978 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="302a0707-59c7-46af-a5dc-cc3c6fa23889" path="/var/lib/kubelet/pods/302a0707-59c7-46af-a5dc-cc3c6fa23889/volumes" Oct 04 10:51:52 crc kubenswrapper[5025]: E1004 10:51:52.648632 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7cbc48db97-z46n4" podUID="b4f50f0c-193b-4e77-a778-2098ba1f756f" Oct 04 10:51:52 crc kubenswrapper[5025]: I1004 10:51:52.678125 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-25rkf" podStartSLOduration=3.678110442 podStartE2EDuration="3.678110442s" podCreationTimestamp="2025-10-04 10:51:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:51:52.656441409 +0000 UTC m=+1041.081408299" watchObservedRunningTime="2025-10-04 10:51:52.678110442 +0000 UTC m=+1041.103077322" Oct 04 10:51:52 crc kubenswrapper[5025]: I1004 10:51:52.718236 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" path="/var/lib/kubelet/pods/e27cdce2-cc3d-4643-b0ce-a7c2f2acf703/volumes" Oct 04 10:51:52 crc kubenswrapper[5025]: I1004 10:51:52.719240 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" event={"ID":"510866f0-02ee-48ef-8219-fdea80015c5a","Type":"ContainerStarted","Data":"3b59104bcfd14dfd14e3f04666af878bf92f5b2ec64c1efe09add8a1ddf9bfcd"} Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.656271 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b4531e3d-3ed8-411c-ae2e-5c0893d250b5","Type":"ContainerStarted","Data":"527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01"} Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.711659 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.757355 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.771285 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68bbfdcf9-mllx9"] Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.827864 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.877422 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5964556bf-mbrwb"] Oct 04 10:51:53 crc kubenswrapper[5025]: E1004 10:51:53.877812 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" containerName="init" Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.877832 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" containerName="init" Oct 04 10:51:53 crc kubenswrapper[5025]: E1004 10:51:53.877848 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="302a0707-59c7-46af-a5dc-cc3c6fa23889" containerName="init" Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.877854 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="302a0707-59c7-46af-a5dc-cc3c6fa23889" containerName="init" Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.878035 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="e27cdce2-cc3d-4643-b0ce-a7c2f2acf703" containerName="init" Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.878056 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="302a0707-59c7-46af-a5dc-cc3c6fa23889" containerName="init" Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.878866 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.913209 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5964556bf-mbrwb"] Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.940425 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a30dd6c-f419-44e2-94e2-8eca192123ae-logs\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.940497 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qz9kd\" (UniqueName: \"kubernetes.io/projected/8a30dd6c-f419-44e2-94e2-8eca192123ae-kube-api-access-qz9kd\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.940540 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-scripts\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.940560 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8a30dd6c-f419-44e2-94e2-8eca192123ae-horizon-secret-key\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:53 crc kubenswrapper[5025]: I1004 10:51:53.940612 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-config-data\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.041850 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-scripts\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.041900 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8a30dd6c-f419-44e2-94e2-8eca192123ae-horizon-secret-key\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.041961 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-config-data\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.042001 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a30dd6c-f419-44e2-94e2-8eca192123ae-logs\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.042061 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qz9kd\" (UniqueName: \"kubernetes.io/projected/8a30dd6c-f419-44e2-94e2-8eca192123ae-kube-api-access-qz9kd\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.042717 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-scripts\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.043185 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a30dd6c-f419-44e2-94e2-8eca192123ae-logs\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.043900 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-config-data\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.067530 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8a30dd6c-f419-44e2-94e2-8eca192123ae-horizon-secret-key\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.076644 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qz9kd\" (UniqueName: \"kubernetes.io/projected/8a30dd6c-f419-44e2-94e2-8eca192123ae-kube-api-access-qz9kd\") pod \"horizon-5964556bf-mbrwb\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.205338 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.668734 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a11a8b2-3ed8-4978-b7ed-298dba8db93d","Type":"ContainerStarted","Data":"fcf161d90c5d4ccb0788643aa9b7e90108c64a1732d7ce5695ee8154061217e0"} Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.669359 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:54 crc kubenswrapper[5025]: I1004 10:51:54.693784 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" podStartSLOduration=5.693767779 podStartE2EDuration="5.693767779s" podCreationTimestamp="2025-10-04 10:51:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:51:54.688888557 +0000 UTC m=+1043.113855427" watchObservedRunningTime="2025-10-04 10:51:54.693767779 +0000 UTC m=+1043.118734659" Oct 04 10:51:55 crc kubenswrapper[5025]: I1004 10:51:55.681578 5025 generic.go:334] "Generic (PLEG): container finished" podID="5b4eee55-362c-41ba-bc6d-4b29c77bddd5" containerID="d6cafb614f4d4c7ccb64f8e00d276df76d61213552f09efaff5591b1e84d47d1" exitCode=0 Oct 04 10:51:55 crc kubenswrapper[5025]: I1004 10:51:55.681668 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsvbc" event={"ID":"5b4eee55-362c-41ba-bc6d-4b29c77bddd5","Type":"ContainerDied","Data":"d6cafb614f4d4c7ccb64f8e00d276df76d61213552f09efaff5591b1e84d47d1"} Oct 04 10:51:56 crc kubenswrapper[5025]: I1004 10:51:56.694320 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a11a8b2-3ed8-4978-b7ed-298dba8db93d","Type":"ContainerStarted","Data":"850ae3b96f8d155eb442fb17daf1ee0170e170d8c9744d6f255b348ebc883891"} Oct 04 10:51:56 crc kubenswrapper[5025]: I1004 10:51:56.694410 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" containerName="glance-log" containerID="cri-o://fcf161d90c5d4ccb0788643aa9b7e90108c64a1732d7ce5695ee8154061217e0" gracePeriod=30 Oct 04 10:51:56 crc kubenswrapper[5025]: I1004 10:51:56.694664 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" containerName="glance-httpd" containerID="cri-o://850ae3b96f8d155eb442fb17daf1ee0170e170d8c9744d6f255b348ebc883891" gracePeriod=30 Oct 04 10:51:56 crc kubenswrapper[5025]: I1004 10:51:56.724957 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.724940109 podStartE2EDuration="7.724940109s" podCreationTimestamp="2025-10-04 10:51:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:51:56.718863321 +0000 UTC m=+1045.143830201" watchObservedRunningTime="2025-10-04 10:51:56.724940109 +0000 UTC m=+1045.149906989" Oct 04 10:51:57 crc kubenswrapper[5025]: I1004 10:51:57.708094 5025 generic.go:334] "Generic (PLEG): container finished" podID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" containerID="850ae3b96f8d155eb442fb17daf1ee0170e170d8c9744d6f255b348ebc883891" exitCode=143 Oct 04 10:51:57 crc kubenswrapper[5025]: I1004 10:51:57.708547 5025 generic.go:334] "Generic (PLEG): container finished" podID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" containerID="fcf161d90c5d4ccb0788643aa9b7e90108c64a1732d7ce5695ee8154061217e0" exitCode=143 Oct 04 10:51:57 crc kubenswrapper[5025]: I1004 10:51:57.708566 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a11a8b2-3ed8-4978-b7ed-298dba8db93d","Type":"ContainerDied","Data":"850ae3b96f8d155eb442fb17daf1ee0170e170d8c9744d6f255b348ebc883891"} Oct 04 10:51:57 crc kubenswrapper[5025]: I1004 10:51:57.708588 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a11a8b2-3ed8-4978-b7ed-298dba8db93d","Type":"ContainerDied","Data":"fcf161d90c5d4ccb0788643aa9b7e90108c64a1732d7ce5695ee8154061217e0"} Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.490328 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.620616 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-combined-ca-bundle\") pod \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.620735 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-credential-keys\") pod \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.620804 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzz7h\" (UniqueName: \"kubernetes.io/projected/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-kube-api-access-kzz7h\") pod \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.620847 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-config-data\") pod \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.620955 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-scripts\") pod \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.621063 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-fernet-keys\") pod \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\" (UID: \"5b4eee55-362c-41ba-bc6d-4b29c77bddd5\") " Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.628848 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-scripts" (OuterVolumeSpecName: "scripts") pod "5b4eee55-362c-41ba-bc6d-4b29c77bddd5" (UID: "5b4eee55-362c-41ba-bc6d-4b29c77bddd5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.628889 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-kube-api-access-kzz7h" (OuterVolumeSpecName: "kube-api-access-kzz7h") pod "5b4eee55-362c-41ba-bc6d-4b29c77bddd5" (UID: "5b4eee55-362c-41ba-bc6d-4b29c77bddd5"). InnerVolumeSpecName "kube-api-access-kzz7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.629410 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5b4eee55-362c-41ba-bc6d-4b29c77bddd5" (UID: "5b4eee55-362c-41ba-bc6d-4b29c77bddd5"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.634178 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5b4eee55-362c-41ba-bc6d-4b29c77bddd5" (UID: "5b4eee55-362c-41ba-bc6d-4b29c77bddd5"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.663069 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b4eee55-362c-41ba-bc6d-4b29c77bddd5" (UID: "5b4eee55-362c-41ba-bc6d-4b29c77bddd5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.663775 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-config-data" (OuterVolumeSpecName: "config-data") pod "5b4eee55-362c-41ba-bc6d-4b29c77bddd5" (UID: "5b4eee55-362c-41ba-bc6d-4b29c77bddd5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.722972 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsvbc" event={"ID":"5b4eee55-362c-41ba-bc6d-4b29c77bddd5","Type":"ContainerDied","Data":"2a7b9c5ae51fa0987c4f55ae4fa95e488699a014eade5e2c8412d7fff1734b1a"} Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.723025 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a7b9c5ae51fa0987c4f55ae4fa95e488699a014eade5e2c8412d7fff1734b1a" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.723123 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsvbc" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.724919 5025 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.724940 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.724949 5025 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.724957 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzz7h\" (UniqueName: \"kubernetes.io/projected/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-kube-api-access-kzz7h\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.724966 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:58 crc kubenswrapper[5025]: I1004 10:51:58.724973 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b4eee55-362c-41ba-bc6d-4b29c77bddd5-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.613581 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hsvbc"] Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.621024 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hsvbc"] Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.696170 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.704841 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dncjm"] Oct 04 10:51:59 crc kubenswrapper[5025]: E1004 10:51:59.705445 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b4eee55-362c-41ba-bc6d-4b29c77bddd5" containerName="keystone-bootstrap" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.705514 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b4eee55-362c-41ba-bc6d-4b29c77bddd5" containerName="keystone-bootstrap" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.705725 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b4eee55-362c-41ba-bc6d-4b29c77bddd5" containerName="keystone-bootstrap" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.706337 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.708194 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9d9vx" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.708354 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.708629 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.708770 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.727269 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dncjm"] Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.775302 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrdkc"] Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.775703 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" podUID="021282fc-71aa-4205-9626-24bd450b837e" containerName="dnsmasq-dns" containerID="cri-o://9c9eb64ba60da9dc8eae7140cb2610c6219b736b7f89ee5e1c7721f1bad4e904" gracePeriod=10 Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.853967 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-credential-keys\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.854913 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-combined-ca-bundle\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.855005 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-scripts\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.855075 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-fernet-keys\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.855347 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-config-data\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.855406 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2frd7\" (UniqueName: \"kubernetes.io/projected/4a4861ec-2e6e-408d-9b42-5ce3248fd640-kube-api-access-2frd7\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.918119 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" podUID="021282fc-71aa-4205-9626-24bd450b837e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: connect: connection refused" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.958341 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-credential-keys\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.958395 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-combined-ca-bundle\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.958421 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-scripts\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.958437 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-fernet-keys\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.958519 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-config-data\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.958548 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2frd7\" (UniqueName: \"kubernetes.io/projected/4a4861ec-2e6e-408d-9b42-5ce3248fd640-kube-api-access-2frd7\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.965374 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-scripts\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.965704 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-fernet-keys\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.966456 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-config-data\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.966490 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-credential-keys\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.980559 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-combined-ca-bundle\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:51:59 crc kubenswrapper[5025]: I1004 10:51:59.984639 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2frd7\" (UniqueName: \"kubernetes.io/projected/4a4861ec-2e6e-408d-9b42-5ce3248fd640-kube-api-access-2frd7\") pod \"keystone-bootstrap-dncjm\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.052836 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.429627 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b4eee55-362c-41ba-bc6d-4b29c77bddd5" path="/var/lib/kubelet/pods/5b4eee55-362c-41ba-bc6d-4b29c77bddd5/volumes" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.430123 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cbc48db97-z46n4"] Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.459187 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5787d4b6b6-fc88x"] Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.460646 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.467645 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.480259 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5787d4b6b6-fc88x"] Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.542723 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5964556bf-mbrwb"] Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.570422 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4337f7-6dc7-4f98-8633-a61b61010693-logs\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.570596 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-scripts\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.570968 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-combined-ca-bundle\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.571138 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-config-data\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.571277 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-secret-key\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.581138 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8tx5\" (UniqueName: \"kubernetes.io/projected/1a4337f7-6dc7-4f98-8633-a61b61010693-kube-api-access-x8tx5\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.581304 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-tls-certs\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.618249 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6559967db4-zfm7d"] Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.619677 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.637474 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6559967db4-zfm7d"] Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.682928 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-scripts\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.682967 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-combined-ca-bundle\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.682991 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-config-data\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.683056 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-secret-key\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.683097 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8tx5\" (UniqueName: \"kubernetes.io/projected/1a4337f7-6dc7-4f98-8633-a61b61010693-kube-api-access-x8tx5\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.683131 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-tls-certs\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.683184 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4337f7-6dc7-4f98-8633-a61b61010693-logs\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.683650 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4337f7-6dc7-4f98-8633-a61b61010693-logs\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.684187 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-scripts\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.688187 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-secret-key\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.691951 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-config-data\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.701601 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-combined-ca-bundle\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.718554 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-tls-certs\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.726112 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8tx5\" (UniqueName: \"kubernetes.io/projected/1a4337f7-6dc7-4f98-8633-a61b61010693-kube-api-access-x8tx5\") pod \"horizon-5787d4b6b6-fc88x\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.780076 5025 generic.go:334] "Generic (PLEG): container finished" podID="021282fc-71aa-4205-9626-24bd450b837e" containerID="9c9eb64ba60da9dc8eae7140cb2610c6219b736b7f89ee5e1c7721f1bad4e904" exitCode=0 Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.780130 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" event={"ID":"021282fc-71aa-4205-9626-24bd450b837e","Type":"ContainerDied","Data":"9c9eb64ba60da9dc8eae7140cb2610c6219b736b7f89ee5e1c7721f1bad4e904"} Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.781571 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.784243 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53944185-08b0-4249-8662-9fdce7cb24cf-logs\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.784429 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53944185-08b0-4249-8662-9fdce7cb24cf-horizon-secret-key\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.784534 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzk4g\" (UniqueName: \"kubernetes.io/projected/53944185-08b0-4249-8662-9fdce7cb24cf-kube-api-access-bzk4g\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.784686 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53944185-08b0-4249-8662-9fdce7cb24cf-combined-ca-bundle\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.784790 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/53944185-08b0-4249-8662-9fdce7cb24cf-horizon-tls-certs\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.784947 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53944185-08b0-4249-8662-9fdce7cb24cf-config-data\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.785052 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53944185-08b0-4249-8662-9fdce7cb24cf-scripts\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.887169 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53944185-08b0-4249-8662-9fdce7cb24cf-horizon-secret-key\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.888125 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzk4g\" (UniqueName: \"kubernetes.io/projected/53944185-08b0-4249-8662-9fdce7cb24cf-kube-api-access-bzk4g\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.888361 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53944185-08b0-4249-8662-9fdce7cb24cf-combined-ca-bundle\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.888575 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/53944185-08b0-4249-8662-9fdce7cb24cf-horizon-tls-certs\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.888723 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53944185-08b0-4249-8662-9fdce7cb24cf-config-data\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.888834 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53944185-08b0-4249-8662-9fdce7cb24cf-scripts\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.888959 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53944185-08b0-4249-8662-9fdce7cb24cf-logs\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.890734 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53944185-08b0-4249-8662-9fdce7cb24cf-logs\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.891170 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53944185-08b0-4249-8662-9fdce7cb24cf-scripts\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.904417 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53944185-08b0-4249-8662-9fdce7cb24cf-horizon-secret-key\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.908102 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/53944185-08b0-4249-8662-9fdce7cb24cf-horizon-tls-certs\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.908329 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53944185-08b0-4249-8662-9fdce7cb24cf-config-data\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.908797 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53944185-08b0-4249-8662-9fdce7cb24cf-combined-ca-bundle\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.911846 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzk4g\" (UniqueName: \"kubernetes.io/projected/53944185-08b0-4249-8662-9fdce7cb24cf-kube-api-access-bzk4g\") pod \"horizon-6559967db4-zfm7d\" (UID: \"53944185-08b0-4249-8662-9fdce7cb24cf\") " pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:00 crc kubenswrapper[5025]: I1004 10:52:00.936126 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:04 crc kubenswrapper[5025]: E1004 10:52:04.164655 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Oct 04 10:52:04 crc kubenswrapper[5025]: E1004 10:52:04.165424 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nc5h679h668hbfh5fdhbh58h9bh647h9fh5bfh584h57hf7h687hc8hchc6h588hfbh67h674h5ffh68bh5f6h64ch99h5bbhbhbh58h695q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vnf6m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-68bbfdcf9-mllx9_openstack(9d2b5229-8605-401c-9321-58b64cd1262a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 10:52:04 crc kubenswrapper[5025]: E1004 10:52:04.175716 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-68bbfdcf9-mllx9" podUID="9d2b5229-8605-401c-9321-58b64cd1262a" Oct 04 10:52:04 crc kubenswrapper[5025]: E1004 10:52:04.667308 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 04 10:52:04 crc kubenswrapper[5025]: E1004 10:52:04.667438 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w7mnr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-sfm5z_openstack(b5d127d5-e426-48fe-bb22-c811de5c801d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 10:52:04 crc kubenswrapper[5025]: E1004 10:52:04.668983 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-sfm5z" podUID="b5d127d5-e426-48fe-bb22-c811de5c801d" Oct 04 10:52:04 crc kubenswrapper[5025]: E1004 10:52:04.884843 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-sfm5z" podUID="b5d127d5-e426-48fe-bb22-c811de5c801d" Oct 04 10:52:04 crc kubenswrapper[5025]: I1004 10:52:04.918686 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" podUID="021282fc-71aa-4205-9626-24bd450b837e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: connect: connection refused" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.372737 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.514072 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-scripts\") pod \"b4f50f0c-193b-4e77-a778-2098ba1f756f\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.514246 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4f50f0c-193b-4e77-a778-2098ba1f756f-logs\") pod \"b4f50f0c-193b-4e77-a778-2098ba1f756f\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.514291 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kxcp\" (UniqueName: \"kubernetes.io/projected/b4f50f0c-193b-4e77-a778-2098ba1f756f-kube-api-access-4kxcp\") pod \"b4f50f0c-193b-4e77-a778-2098ba1f756f\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.514357 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-config-data\") pod \"b4f50f0c-193b-4e77-a778-2098ba1f756f\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.514473 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b4f50f0c-193b-4e77-a778-2098ba1f756f-horizon-secret-key\") pod \"b4f50f0c-193b-4e77-a778-2098ba1f756f\" (UID: \"b4f50f0c-193b-4e77-a778-2098ba1f756f\") " Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.514543 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4f50f0c-193b-4e77-a778-2098ba1f756f-logs" (OuterVolumeSpecName: "logs") pod "b4f50f0c-193b-4e77-a778-2098ba1f756f" (UID: "b4f50f0c-193b-4e77-a778-2098ba1f756f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.514673 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-scripts" (OuterVolumeSpecName: "scripts") pod "b4f50f0c-193b-4e77-a778-2098ba1f756f" (UID: "b4f50f0c-193b-4e77-a778-2098ba1f756f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.515312 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4f50f0c-193b-4e77-a778-2098ba1f756f-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.515338 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.515717 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-config-data" (OuterVolumeSpecName: "config-data") pod "b4f50f0c-193b-4e77-a778-2098ba1f756f" (UID: "b4f50f0c-193b-4e77-a778-2098ba1f756f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.520650 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f50f0c-193b-4e77-a778-2098ba1f756f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "b4f50f0c-193b-4e77-a778-2098ba1f756f" (UID: "b4f50f0c-193b-4e77-a778-2098ba1f756f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.521165 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4f50f0c-193b-4e77-a778-2098ba1f756f-kube-api-access-4kxcp" (OuterVolumeSpecName: "kube-api-access-4kxcp") pod "b4f50f0c-193b-4e77-a778-2098ba1f756f" (UID: "b4f50f0c-193b-4e77-a778-2098ba1f756f"). InnerVolumeSpecName "kube-api-access-4kxcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.617645 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b4f50f0c-193b-4e77-a778-2098ba1f756f-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.617683 5025 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b4f50f0c-193b-4e77-a778-2098ba1f756f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.617699 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kxcp\" (UniqueName: \"kubernetes.io/projected/b4f50f0c-193b-4e77-a778-2098ba1f756f-kube-api-access-4kxcp\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.843496 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7cbc48db97-z46n4" event={"ID":"b4f50f0c-193b-4e77-a778-2098ba1f756f","Type":"ContainerDied","Data":"5a8772e3f6c17cbcb214c9a46fe8b04870d133a173aea59e108275bc40ddf22b"} Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.843527 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7cbc48db97-z46n4" Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.913101 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7cbc48db97-z46n4"] Oct 04 10:52:06 crc kubenswrapper[5025]: I1004 10:52:06.921378 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7cbc48db97-z46n4"] Oct 04 10:52:08 crc kubenswrapper[5025]: I1004 10:52:08.422078 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4f50f0c-193b-4e77-a778-2098ba1f756f" path="/var/lib/kubelet/pods/b4f50f0c-193b-4e77-a778-2098ba1f756f/volumes" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.719063 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.727310 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.869898 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnf6m\" (UniqueName: \"kubernetes.io/projected/9d2b5229-8605-401c-9321-58b64cd1262a-kube-api-access-vnf6m\") pod \"9d2b5229-8605-401c-9321-58b64cd1262a\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870244 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-scripts\") pod \"9d2b5229-8605-401c-9321-58b64cd1262a\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870344 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-scripts\") pod \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870373 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph9rb\" (UniqueName: \"kubernetes.io/projected/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-kube-api-access-ph9rb\") pod \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870404 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870424 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d2b5229-8605-401c-9321-58b64cd1262a-horizon-secret-key\") pod \"9d2b5229-8605-401c-9321-58b64cd1262a\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870463 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-httpd-run\") pod \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870506 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-combined-ca-bundle\") pod \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870543 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-logs\") pod \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870569 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-config-data\") pod \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\" (UID: \"7a11a8b2-3ed8-4978-b7ed-298dba8db93d\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870599 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2b5229-8605-401c-9321-58b64cd1262a-logs\") pod \"9d2b5229-8605-401c-9321-58b64cd1262a\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870647 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-config-data\") pod \"9d2b5229-8605-401c-9321-58b64cd1262a\" (UID: \"9d2b5229-8605-401c-9321-58b64cd1262a\") " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.870764 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-scripts" (OuterVolumeSpecName: "scripts") pod "9d2b5229-8605-401c-9321-58b64cd1262a" (UID: "9d2b5229-8605-401c-9321-58b64cd1262a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.871376 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7a11a8b2-3ed8-4978-b7ed-298dba8db93d" (UID: "7a11a8b2-3ed8-4978-b7ed-298dba8db93d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.871716 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d2b5229-8605-401c-9321-58b64cd1262a-logs" (OuterVolumeSpecName: "logs") pod "9d2b5229-8605-401c-9321-58b64cd1262a" (UID: "9d2b5229-8605-401c-9321-58b64cd1262a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.871903 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-logs" (OuterVolumeSpecName: "logs") pod "7a11a8b2-3ed8-4978-b7ed-298dba8db93d" (UID: "7a11a8b2-3ed8-4978-b7ed-298dba8db93d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.871965 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-config-data" (OuterVolumeSpecName: "config-data") pod "9d2b5229-8605-401c-9321-58b64cd1262a" (UID: "9d2b5229-8605-401c-9321-58b64cd1262a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.872039 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.876675 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d2b5229-8605-401c-9321-58b64cd1262a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9d2b5229-8605-401c-9321-58b64cd1262a" (UID: "9d2b5229-8605-401c-9321-58b64cd1262a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.876701 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-scripts" (OuterVolumeSpecName: "scripts") pod "7a11a8b2-3ed8-4978-b7ed-298dba8db93d" (UID: "7a11a8b2-3ed8-4978-b7ed-298dba8db93d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.877137 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "7a11a8b2-3ed8-4978-b7ed-298dba8db93d" (UID: "7a11a8b2-3ed8-4978-b7ed-298dba8db93d"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.877297 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d2b5229-8605-401c-9321-58b64cd1262a-kube-api-access-vnf6m" (OuterVolumeSpecName: "kube-api-access-vnf6m") pod "9d2b5229-8605-401c-9321-58b64cd1262a" (UID: "9d2b5229-8605-401c-9321-58b64cd1262a"). InnerVolumeSpecName "kube-api-access-vnf6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.878787 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-kube-api-access-ph9rb" (OuterVolumeSpecName: "kube-api-access-ph9rb") pod "7a11a8b2-3ed8-4978-b7ed-298dba8db93d" (UID: "7a11a8b2-3ed8-4978-b7ed-298dba8db93d"). InnerVolumeSpecName "kube-api-access-ph9rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.896331 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-68bbfdcf9-mllx9" event={"ID":"9d2b5229-8605-401c-9321-58b64cd1262a","Type":"ContainerDied","Data":"9fe90c4608846605874c26c6faa0bfdc97b95c511d6c3d10ea36d872ceed2bf7"} Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.896476 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-68bbfdcf9-mllx9" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.902175 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7a11a8b2-3ed8-4978-b7ed-298dba8db93d","Type":"ContainerDied","Data":"843a03026ef2521cbfd72cf87dc1e0472fd7f53fed109a3080292f49e4fd6c8d"} Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.902217 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.902240 5025 scope.go:117] "RemoveContainer" containerID="850ae3b96f8d155eb442fb17daf1ee0170e170d8c9744d6f255b348ebc883891" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.906382 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a11a8b2-3ed8-4978-b7ed-298dba8db93d" (UID: "7a11a8b2-3ed8-4978-b7ed-298dba8db93d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.926777 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-config-data" (OuterVolumeSpecName: "config-data") pod "7a11a8b2-3ed8-4978-b7ed-298dba8db93d" (UID: "7a11a8b2-3ed8-4978-b7ed-298dba8db93d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973234 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973265 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973279 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973291 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d2b5229-8605-401c-9321-58b64cd1262a-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973302 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9d2b5229-8605-401c-9321-58b64cd1262a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973315 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnf6m\" (UniqueName: \"kubernetes.io/projected/9d2b5229-8605-401c-9321-58b64cd1262a-kube-api-access-vnf6m\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973328 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973339 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph9rb\" (UniqueName: \"kubernetes.io/projected/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-kube-api-access-ph9rb\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973367 5025 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973378 5025 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9d2b5229-8605-401c-9321-58b64cd1262a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.973390 5025 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7a11a8b2-3ed8-4978-b7ed-298dba8db93d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:12 crc kubenswrapper[5025]: I1004 10:52:12.997418 5025 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.024269 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-68bbfdcf9-mllx9"] Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.029864 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-68bbfdcf9-mllx9"] Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.075268 5025 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:13 crc kubenswrapper[5025]: E1004 10:52:13.089044 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Oct 04 10:52:13 crc kubenswrapper[5025]: E1004 10:52:13.089212 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nbfh5f9hb7h56fh547h5dh96h64bh569h9h98h5d7h5dch557h564h5c7hc7h558h689hd6h5dh58fh659h5bch695h9bh58ch566h556h8fh679h56fq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wbbdh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(03dee37f-5b63-4b21-b401-f05b07d3b1af): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.190582 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.256760 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.287138 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.302104 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:52:13 crc kubenswrapper[5025]: E1004 10:52:13.302892 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="021282fc-71aa-4205-9626-24bd450b837e" containerName="init" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.302911 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="021282fc-71aa-4205-9626-24bd450b837e" containerName="init" Oct 04 10:52:13 crc kubenswrapper[5025]: E1004 10:52:13.302935 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="021282fc-71aa-4205-9626-24bd450b837e" containerName="dnsmasq-dns" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.302943 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="021282fc-71aa-4205-9626-24bd450b837e" containerName="dnsmasq-dns" Oct 04 10:52:13 crc kubenswrapper[5025]: E1004 10:52:13.302985 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" containerName="glance-httpd" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.302994 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" containerName="glance-httpd" Oct 04 10:52:13 crc kubenswrapper[5025]: E1004 10:52:13.303091 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" containerName="glance-log" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.303108 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" containerName="glance-log" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.303592 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="021282fc-71aa-4205-9626-24bd450b837e" containerName="dnsmasq-dns" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.303624 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" containerName="glance-log" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.303654 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" containerName="glance-httpd" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.310890 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.312265 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.314746 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.317327 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.394887 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f95qq\" (UniqueName: \"kubernetes.io/projected/021282fc-71aa-4205-9626-24bd450b837e-kube-api-access-f95qq\") pod \"021282fc-71aa-4205-9626-24bd450b837e\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.394944 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-sb\") pod \"021282fc-71aa-4205-9626-24bd450b837e\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.395026 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-nb\") pod \"021282fc-71aa-4205-9626-24bd450b837e\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.395114 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-config\") pod \"021282fc-71aa-4205-9626-24bd450b837e\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.395161 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-svc\") pod \"021282fc-71aa-4205-9626-24bd450b837e\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.395183 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-swift-storage-0\") pod \"021282fc-71aa-4205-9626-24bd450b837e\" (UID: \"021282fc-71aa-4205-9626-24bd450b837e\") " Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.401145 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/021282fc-71aa-4205-9626-24bd450b837e-kube-api-access-f95qq" (OuterVolumeSpecName: "kube-api-access-f95qq") pod "021282fc-71aa-4205-9626-24bd450b837e" (UID: "021282fc-71aa-4205-9626-24bd450b837e"). InnerVolumeSpecName "kube-api-access-f95qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.435180 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "021282fc-71aa-4205-9626-24bd450b837e" (UID: "021282fc-71aa-4205-9626-24bd450b837e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.450152 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "021282fc-71aa-4205-9626-24bd450b837e" (UID: "021282fc-71aa-4205-9626-24bd450b837e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.454227 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "021282fc-71aa-4205-9626-24bd450b837e" (UID: "021282fc-71aa-4205-9626-24bd450b837e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.458995 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-config" (OuterVolumeSpecName: "config") pod "021282fc-71aa-4205-9626-24bd450b837e" (UID: "021282fc-71aa-4205-9626-24bd450b837e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.468618 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "021282fc-71aa-4205-9626-24bd450b837e" (UID: "021282fc-71aa-4205-9626-24bd450b837e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.497165 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.497230 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.497384 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49n6s\" (UniqueName: \"kubernetes.io/projected/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-kube-api-access-49n6s\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.497523 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.497788 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-logs\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.497859 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.497921 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.498105 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.498235 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.498258 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.498270 5025 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.498285 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f95qq\" (UniqueName: \"kubernetes.io/projected/021282fc-71aa-4205-9626-24bd450b837e-kube-api-access-f95qq\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.498297 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.498332 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/021282fc-71aa-4205-9626-24bd450b837e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.559945 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5964556bf-mbrwb"] Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.600125 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.600541 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.600604 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.600658 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.600689 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.600987 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.601487 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49n6s\" (UniqueName: \"kubernetes.io/projected/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-kube-api-access-49n6s\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.601542 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.601694 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-logs\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.602172 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-logs\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.602554 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.604008 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.604283 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.605279 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.606290 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.621859 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49n6s\" (UniqueName: \"kubernetes.io/projected/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-kube-api-access-49n6s\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.637352 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.648349 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.911908 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" event={"ID":"021282fc-71aa-4205-9626-24bd450b837e","Type":"ContainerDied","Data":"0dc35a110c597c215407d966dcb34fef1b56bfc78e1fa55941a94dde7cb367d3"} Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.912002 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.954603 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrdkc"] Oct 04 10:52:13 crc kubenswrapper[5025]: I1004 10:52:13.960576 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-nrdkc"] Oct 04 10:52:14 crc kubenswrapper[5025]: I1004 10:52:14.423005 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="021282fc-71aa-4205-9626-24bd450b837e" path="/var/lib/kubelet/pods/021282fc-71aa-4205-9626-24bd450b837e/volumes" Oct 04 10:52:14 crc kubenswrapper[5025]: I1004 10:52:14.424357 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a11a8b2-3ed8-4978-b7ed-298dba8db93d" path="/var/lib/kubelet/pods/7a11a8b2-3ed8-4978-b7ed-298dba8db93d/volumes" Oct 04 10:52:14 crc kubenswrapper[5025]: I1004 10:52:14.425827 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d2b5229-8605-401c-9321-58b64cd1262a" path="/var/lib/kubelet/pods/9d2b5229-8605-401c-9321-58b64cd1262a/volumes" Oct 04 10:52:14 crc kubenswrapper[5025]: W1004 10:52:14.520425 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a30dd6c_f419_44e2_94e2_8eca192123ae.slice/crio-cc8dc546c2918adef5c44bb5948375bbb362c06046a5ecc529bd37e6197ee9cf WatchSource:0}: Error finding container cc8dc546c2918adef5c44bb5948375bbb362c06046a5ecc529bd37e6197ee9cf: Status 404 returned error can't find the container with id cc8dc546c2918adef5c44bb5948375bbb362c06046a5ecc529bd37e6197ee9cf Oct 04 10:52:14 crc kubenswrapper[5025]: I1004 10:52:14.544659 5025 scope.go:117] "RemoveContainer" containerID="fcf161d90c5d4ccb0788643aa9b7e90108c64a1732d7ce5695ee8154061217e0" Oct 04 10:52:14 crc kubenswrapper[5025]: E1004 10:52:14.560872 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 04 10:52:14 crc kubenswrapper[5025]: E1004 10:52:14.560990 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-68hzg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-xtspb_openstack(694e381e-49b8-49f1-a5ae-7aa3b72d8fee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 10:52:14 crc kubenswrapper[5025]: E1004 10:52:14.562212 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-xtspb" podUID="694e381e-49b8-49f1-a5ae-7aa3b72d8fee" Oct 04 10:52:14 crc kubenswrapper[5025]: I1004 10:52:14.676665 5025 scope.go:117] "RemoveContainer" containerID="9c9eb64ba60da9dc8eae7140cb2610c6219b736b7f89ee5e1c7721f1bad4e904" Oct 04 10:52:14 crc kubenswrapper[5025]: I1004 10:52:14.759207 5025 scope.go:117] "RemoveContainer" containerID="0bf38c469326095aa149c09982e2dfbee19aa502fe7dbda60dec10caeefe4abe" Oct 04 10:52:14 crc kubenswrapper[5025]: I1004 10:52:14.917587 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-nrdkc" podUID="021282fc-71aa-4205-9626-24bd450b837e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: i/o timeout" Oct 04 10:52:14 crc kubenswrapper[5025]: I1004 10:52:14.923908 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g94ln" event={"ID":"565cc52d-cdcf-4d59-82b5-3d71804b727a","Type":"ContainerStarted","Data":"52df45af5dc37cc05f6bc5aec470bfdecd9f2f4c802694476feefddd6e716cf1"} Oct 04 10:52:14 crc kubenswrapper[5025]: I1004 10:52:14.928684 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5964556bf-mbrwb" event={"ID":"8a30dd6c-f419-44e2-94e2-8eca192123ae","Type":"ContainerStarted","Data":"cc8dc546c2918adef5c44bb5948375bbb362c06046a5ecc529bd37e6197ee9cf"} Oct 04 10:52:14 crc kubenswrapper[5025]: E1004 10:52:14.930596 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-xtspb" podUID="694e381e-49b8-49f1-a5ae-7aa3b72d8fee" Oct 04 10:52:14 crc kubenswrapper[5025]: I1004 10:52:14.940306 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-g94ln" podStartSLOduration=3.055109985 podStartE2EDuration="25.940291505s" podCreationTimestamp="2025-10-04 10:51:49 +0000 UTC" firstStartedPulling="2025-10-04 10:51:50.227098573 +0000 UTC m=+1038.652065443" lastFinishedPulling="2025-10-04 10:52:13.112280083 +0000 UTC m=+1061.537246963" observedRunningTime="2025-10-04 10:52:14.938910855 +0000 UTC m=+1063.363877755" watchObservedRunningTime="2025-10-04 10:52:14.940291505 +0000 UTC m=+1063.365258375" Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.021279 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6559967db4-zfm7d"] Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.124084 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5787d4b6b6-fc88x"] Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.132944 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dncjm"] Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.413948 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:52:15 crc kubenswrapper[5025]: W1004 10:52:15.633155 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ef75483_c407_4ba6_8cfe_9b7a663ce7f5.slice/crio-1ccf6b52cc6995544a2732356a6ab15d6ec29829771f896f0c034fe9d6926570 WatchSource:0}: Error finding container 1ccf6b52cc6995544a2732356a6ab15d6ec29829771f896f0c034fe9d6926570: Status 404 returned error can't find the container with id 1ccf6b52cc6995544a2732356a6ab15d6ec29829771f896f0c034fe9d6926570 Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.939270 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b4531e3d-3ed8-411c-ae2e-5c0893d250b5","Type":"ContainerStarted","Data":"dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae"} Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.939756 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" containerName="glance-log" containerID="cri-o://527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01" gracePeriod=30 Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.940000 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" containerName="glance-httpd" containerID="cri-o://dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae" gracePeriod=30 Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.951355 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5787d4b6b6-fc88x" event={"ID":"1a4337f7-6dc7-4f98-8633-a61b61010693","Type":"ContainerStarted","Data":"16fdc70e3b1d7f1e5bbf6f4cd085a28452869621c0a13d1016341a4f70694f64"} Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.951406 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5787d4b6b6-fc88x" event={"ID":"1a4337f7-6dc7-4f98-8633-a61b61010693","Type":"ContainerStarted","Data":"db5d128e43585adecdb5dcfa84b1cd7d2595aafa9047d23a2c19faf08d09a728"} Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.957516 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5","Type":"ContainerStarted","Data":"1ccf6b52cc6995544a2732356a6ab15d6ec29829771f896f0c034fe9d6926570"} Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.959541 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5964556bf-mbrwb" event={"ID":"8a30dd6c-f419-44e2-94e2-8eca192123ae","Type":"ContainerStarted","Data":"0a9d774c273867c86c28003916c2dd1cd5282a66d98716a6a4b6bc2026604acd"} Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.963061 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6559967db4-zfm7d" event={"ID":"53944185-08b0-4249-8662-9fdce7cb24cf","Type":"ContainerStarted","Data":"a9b8782c20348f10d6f663dedb1ad019f1572ccdb249fcdad9a3280f7f2802cb"} Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.963106 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6559967db4-zfm7d" event={"ID":"53944185-08b0-4249-8662-9fdce7cb24cf","Type":"ContainerStarted","Data":"6a18f3f73af4bca577d51a92edb8d177944597613f9eaf77c3af179a16208649"} Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.971086 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=26.97106775 podStartE2EDuration="26.97106775s" podCreationTimestamp="2025-10-04 10:51:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:15.96662425 +0000 UTC m=+1064.391591120" watchObservedRunningTime="2025-10-04 10:52:15.97106775 +0000 UTC m=+1064.396034650" Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.998266 5025 generic.go:334] "Generic (PLEG): container finished" podID="8e307d63-f22a-4e97-9385-773e9ce7f4fd" containerID="6bb3b61562fd3fec493b942be4e930f472a30c9bb280ebef518fa016041ed7b7" exitCode=0 Oct 04 10:52:15 crc kubenswrapper[5025]: I1004 10:52:15.998378 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-25rkf" event={"ID":"8e307d63-f22a-4e97-9385-773e9ce7f4fd","Type":"ContainerDied","Data":"6bb3b61562fd3fec493b942be4e930f472a30c9bb280ebef518fa016041ed7b7"} Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.003417 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dncjm" event={"ID":"4a4861ec-2e6e-408d-9b42-5ce3248fd640","Type":"ContainerStarted","Data":"27f95935eab866d09895b9f363cc4d687040656a7e9cf672deb677ef1396860a"} Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.003449 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dncjm" event={"ID":"4a4861ec-2e6e-408d-9b42-5ce3248fd640","Type":"ContainerStarted","Data":"d8342b5433e0d1474ee87088d4056cba0fbcdbe10f7c3566f6f75fd061fd7d1b"} Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.036919 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dncjm" podStartSLOduration=17.036904014 podStartE2EDuration="17.036904014s" podCreationTimestamp="2025-10-04 10:51:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:16.036626825 +0000 UTC m=+1064.461593705" watchObservedRunningTime="2025-10-04 10:52:16.036904014 +0000 UTC m=+1064.461870894" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.565658 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.662937 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-config-data\") pod \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.663214 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-combined-ca-bundle\") pod \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.663255 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-httpd-run\") pod \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.663326 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.663410 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k86p\" (UniqueName: \"kubernetes.io/projected/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-kube-api-access-7k86p\") pod \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.663451 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-logs\") pod \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.663514 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-scripts\") pod \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\" (UID: \"b4531e3d-3ed8-411c-ae2e-5c0893d250b5\") " Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.664602 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-logs" (OuterVolumeSpecName: "logs") pod "b4531e3d-3ed8-411c-ae2e-5c0893d250b5" (UID: "b4531e3d-3ed8-411c-ae2e-5c0893d250b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.664976 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b4531e3d-3ed8-411c-ae2e-5c0893d250b5" (UID: "b4531e3d-3ed8-411c-ae2e-5c0893d250b5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.669394 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "b4531e3d-3ed8-411c-ae2e-5c0893d250b5" (UID: "b4531e3d-3ed8-411c-ae2e-5c0893d250b5"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.669415 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-kube-api-access-7k86p" (OuterVolumeSpecName: "kube-api-access-7k86p") pod "b4531e3d-3ed8-411c-ae2e-5c0893d250b5" (UID: "b4531e3d-3ed8-411c-ae2e-5c0893d250b5"). InnerVolumeSpecName "kube-api-access-7k86p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.670435 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-scripts" (OuterVolumeSpecName: "scripts") pod "b4531e3d-3ed8-411c-ae2e-5c0893d250b5" (UID: "b4531e3d-3ed8-411c-ae2e-5c0893d250b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.700129 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4531e3d-3ed8-411c-ae2e-5c0893d250b5" (UID: "b4531e3d-3ed8-411c-ae2e-5c0893d250b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.716000 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-config-data" (OuterVolumeSpecName: "config-data") pod "b4531e3d-3ed8-411c-ae2e-5c0893d250b5" (UID: "b4531e3d-3ed8-411c-ae2e-5c0893d250b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.764956 5025 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.764992 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k86p\" (UniqueName: \"kubernetes.io/projected/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-kube-api-access-7k86p\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.765003 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.765027 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.765037 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.765045 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.765053 5025 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4531e3d-3ed8-411c-ae2e-5c0893d250b5-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.781114 5025 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 10:52:16 crc kubenswrapper[5025]: I1004 10:52:16.878623 5025 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.016283 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6559967db4-zfm7d" event={"ID":"53944185-08b0-4249-8662-9fdce7cb24cf","Type":"ContainerStarted","Data":"c2a5aa45f2eb46777bb924b8fb577fd10ebc7792e6aff7c0169bdf49da815f16"} Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.022199 5025 generic.go:334] "Generic (PLEG): container finished" podID="565cc52d-cdcf-4d59-82b5-3d71804b727a" containerID="52df45af5dc37cc05f6bc5aec470bfdecd9f2f4c802694476feefddd6e716cf1" exitCode=0 Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.022320 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g94ln" event={"ID":"565cc52d-cdcf-4d59-82b5-3d71804b727a","Type":"ContainerDied","Data":"52df45af5dc37cc05f6bc5aec470bfdecd9f2f4c802694476feefddd6e716cf1"} Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.027294 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03dee37f-5b63-4b21-b401-f05b07d3b1af","Type":"ContainerStarted","Data":"872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33"} Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.030400 5025 generic.go:334] "Generic (PLEG): container finished" podID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" containerID="dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae" exitCode=0 Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.030456 5025 generic.go:334] "Generic (PLEG): container finished" podID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" containerID="527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01" exitCode=143 Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.030535 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b4531e3d-3ed8-411c-ae2e-5c0893d250b5","Type":"ContainerDied","Data":"dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae"} Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.030607 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b4531e3d-3ed8-411c-ae2e-5c0893d250b5","Type":"ContainerDied","Data":"527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01"} Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.030629 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b4531e3d-3ed8-411c-ae2e-5c0893d250b5","Type":"ContainerDied","Data":"479a50bb495ab54164d313abef8cc81523b09af4b7949fa6a7c7aa680a2b07e8"} Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.030726 5025 scope.go:117] "RemoveContainer" containerID="dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.030986 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.042327 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5787d4b6b6-fc88x" event={"ID":"1a4337f7-6dc7-4f98-8633-a61b61010693","Type":"ContainerStarted","Data":"64a255b2f77c316007bca29fbea4431bb259ddb5e8fe92d31a3ac65b6b471726"} Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.049388 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5","Type":"ContainerStarted","Data":"9728653a1a8077c2d9e3b1c196c78e1f5847d3dc50e63f5e40634ec0f07be6d6"} Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.052313 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5964556bf-mbrwb" podUID="8a30dd6c-f419-44e2-94e2-8eca192123ae" containerName="horizon-log" containerID="cri-o://0a9d774c273867c86c28003916c2dd1cd5282a66d98716a6a4b6bc2026604acd" gracePeriod=30 Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.052365 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5964556bf-mbrwb" event={"ID":"8a30dd6c-f419-44e2-94e2-8eca192123ae","Type":"ContainerStarted","Data":"d69338dcf54457fcdb1b6f100e8efeb9df6797e7ae70f2ffd70bbd09b4194103"} Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.052498 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5964556bf-mbrwb" podUID="8a30dd6c-f419-44e2-94e2-8eca192123ae" containerName="horizon" containerID="cri-o://d69338dcf54457fcdb1b6f100e8efeb9df6797e7ae70f2ffd70bbd09b4194103" gracePeriod=30 Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.060058 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6559967db4-zfm7d" podStartSLOduration=17.060031904 podStartE2EDuration="17.060031904s" podCreationTimestamp="2025-10-04 10:52:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:17.039048681 +0000 UTC m=+1065.464015581" watchObservedRunningTime="2025-10-04 10:52:17.060031904 +0000 UTC m=+1065.484998784" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.083394 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5964556bf-mbrwb" podStartSLOduration=23.620437558 podStartE2EDuration="24.083376467s" podCreationTimestamp="2025-10-04 10:51:53 +0000 UTC" firstStartedPulling="2025-10-04 10:52:14.525209775 +0000 UTC m=+1062.950176655" lastFinishedPulling="2025-10-04 10:52:14.988148684 +0000 UTC m=+1063.413115564" observedRunningTime="2025-10-04 10:52:17.07287755 +0000 UTC m=+1065.497844440" watchObservedRunningTime="2025-10-04 10:52:17.083376467 +0000 UTC m=+1065.508343347" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.084606 5025 scope.go:117] "RemoveContainer" containerID="527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.111811 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5787d4b6b6-fc88x" podStartSLOduration=17.111787187 podStartE2EDuration="17.111787187s" podCreationTimestamp="2025-10-04 10:52:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:17.09923281 +0000 UTC m=+1065.524199690" watchObservedRunningTime="2025-10-04 10:52:17.111787187 +0000 UTC m=+1065.536754067" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.126091 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.128976 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.135617 5025 scope.go:117] "RemoveContainer" containerID="dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae" Oct 04 10:52:17 crc kubenswrapper[5025]: E1004 10:52:17.136130 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae\": container with ID starting with dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae not found: ID does not exist" containerID="dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.136158 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae"} err="failed to get container status \"dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae\": rpc error: code = NotFound desc = could not find container \"dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae\": container with ID starting with dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae not found: ID does not exist" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.136181 5025 scope.go:117] "RemoveContainer" containerID="527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01" Oct 04 10:52:17 crc kubenswrapper[5025]: E1004 10:52:17.136393 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01\": container with ID starting with 527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01 not found: ID does not exist" containerID="527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.136412 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01"} err="failed to get container status \"527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01\": rpc error: code = NotFound desc = could not find container \"527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01\": container with ID starting with 527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01 not found: ID does not exist" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.136424 5025 scope.go:117] "RemoveContainer" containerID="dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.136628 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae"} err="failed to get container status \"dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae\": rpc error: code = NotFound desc = could not find container \"dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae\": container with ID starting with dd30316904469853513721c13167710e04ecb301a743563ef87e4c9033d72fae not found: ID does not exist" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.136646 5025 scope.go:117] "RemoveContainer" containerID="527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.136789 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01"} err="failed to get container status \"527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01\": rpc error: code = NotFound desc = could not find container \"527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01\": container with ID starting with 527f9067ed94dfb880d8aacd55d2d079211b96897daba1441bab78f9bb1d9f01 not found: ID does not exist" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.162614 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:52:17 crc kubenswrapper[5025]: E1004 10:52:17.163169 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" containerName="glance-log" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.163196 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" containerName="glance-log" Oct 04 10:52:17 crc kubenswrapper[5025]: E1004 10:52:17.163229 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" containerName="glance-httpd" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.163241 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" containerName="glance-httpd" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.163495 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" containerName="glance-log" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.163543 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" containerName="glance-httpd" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.164841 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.179212 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.181041 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.183672 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.289330 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-logs\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.289395 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49hr4\" (UniqueName: \"kubernetes.io/projected/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-kube-api-access-49hr4\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.289439 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.289522 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-scripts\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.289609 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-config-data\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.289637 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.289674 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.289856 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.392195 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-logs\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.392246 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49hr4\" (UniqueName: \"kubernetes.io/projected/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-kube-api-access-49hr4\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.392273 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.392329 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-scripts\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.392376 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-config-data\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.392404 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.392437 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.392477 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.392675 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-logs\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.392918 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.393262 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.396656 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.406600 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-config-data\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.411539 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-scripts\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.418390 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.425439 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49hr4\" (UniqueName: \"kubernetes.io/projected/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-kube-api-access-49hr4\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.437467 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-25rkf" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.447280 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.512392 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.594931 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rs59\" (UniqueName: \"kubernetes.io/projected/8e307d63-f22a-4e97-9385-773e9ce7f4fd-kube-api-access-4rs59\") pod \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.595155 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-config\") pod \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.595240 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-combined-ca-bundle\") pod \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\" (UID: \"8e307d63-f22a-4e97-9385-773e9ce7f4fd\") " Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.606367 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e307d63-f22a-4e97-9385-773e9ce7f4fd-kube-api-access-4rs59" (OuterVolumeSpecName: "kube-api-access-4rs59") pod "8e307d63-f22a-4e97-9385-773e9ce7f4fd" (UID: "8e307d63-f22a-4e97-9385-773e9ce7f4fd"). InnerVolumeSpecName "kube-api-access-4rs59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.631372 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e307d63-f22a-4e97-9385-773e9ce7f4fd" (UID: "8e307d63-f22a-4e97-9385-773e9ce7f4fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.634385 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-config" (OuterVolumeSpecName: "config") pod "8e307d63-f22a-4e97-9385-773e9ce7f4fd" (UID: "8e307d63-f22a-4e97-9385-773e9ce7f4fd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.698389 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.698421 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e307d63-f22a-4e97-9385-773e9ce7f4fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:17 crc kubenswrapper[5025]: I1004 10:52:17.698436 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rs59\" (UniqueName: \"kubernetes.io/projected/8e307d63-f22a-4e97-9385-773e9ce7f4fd-kube-api-access-4rs59\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.079562 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-25rkf" event={"ID":"8e307d63-f22a-4e97-9385-773e9ce7f4fd","Type":"ContainerDied","Data":"0070e3e8dda26726f80f094b06496edcff93989e8c55dcf61cb0be944e4e59a3"} Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.079909 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0070e3e8dda26726f80f094b06496edcff93989e8c55dcf61cb0be944e4e59a3" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.080441 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-25rkf" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.132207 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5","Type":"ContainerStarted","Data":"cfe3bfe62fe9e218a73048a8aef6b8aede522fd5277af0300f8f2796b80ac10b"} Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.260379 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.262923 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.262906037 podStartE2EDuration="5.262906037s" podCreationTimestamp="2025-10-04 10:52:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:18.236210407 +0000 UTC m=+1066.661177287" watchObservedRunningTime="2025-10-04 10:52:18.262906037 +0000 UTC m=+1066.687872917" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.298873 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-c6gbx"] Oct 04 10:52:18 crc kubenswrapper[5025]: E1004 10:52:18.299295 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e307d63-f22a-4e97-9385-773e9ce7f4fd" containerName="neutron-db-sync" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.299316 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e307d63-f22a-4e97-9385-773e9ce7f4fd" containerName="neutron-db-sync" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.299535 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e307d63-f22a-4e97-9385-773e9ce7f4fd" containerName="neutron-db-sync" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.300746 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.325827 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-c6gbx"] Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.412483 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.412552 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.412590 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.412925 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cncjg\" (UniqueName: \"kubernetes.io/projected/d87fc908-36a3-4364-a4e4-85bb1abb660a-kube-api-access-cncjg\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.412967 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.413007 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-config\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.426851 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4531e3d-3ed8-411c-ae2e-5c0893d250b5" path="/var/lib/kubelet/pods/b4531e3d-3ed8-411c-ae2e-5c0893d250b5/volumes" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.435504 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c7cc4784d-fsjlq"] Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.436946 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.457542 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.457781 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.458006 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-d8fx9" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.458434 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.478180 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c7cc4784d-fsjlq"] Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.517905 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.517980 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-combined-ca-bundle\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.518007 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfv7n\" (UniqueName: \"kubernetes.io/projected/6622de76-40c3-434a-a87d-127e3d50fd4e-kube-api-access-zfv7n\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.518036 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-httpd-config\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.518083 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.518116 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.518133 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-ovndb-tls-certs\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.518156 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-config\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.518207 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cncjg\" (UniqueName: \"kubernetes.io/projected/d87fc908-36a3-4364-a4e4-85bb1abb660a-kube-api-access-cncjg\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.518234 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.518268 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-config\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.522483 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-config\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.523105 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.525180 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.525696 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.526412 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.560072 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cncjg\" (UniqueName: \"kubernetes.io/projected/d87fc908-36a3-4364-a4e4-85bb1abb660a-kube-api-access-cncjg\") pod \"dnsmasq-dns-5ccc5c4795-c6gbx\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.620162 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-combined-ca-bundle\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.620207 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfv7n\" (UniqueName: \"kubernetes.io/projected/6622de76-40c3-434a-a87d-127e3d50fd4e-kube-api-access-zfv7n\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.620242 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-httpd-config\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.620295 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-ovndb-tls-certs\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.620338 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-config\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.625436 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-config\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.628035 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-combined-ca-bundle\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.628323 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-httpd-config\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.636922 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfv7n\" (UniqueName: \"kubernetes.io/projected/6622de76-40c3-434a-a87d-127e3d50fd4e-kube-api-access-zfv7n\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.645520 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-ovndb-tls-certs\") pod \"neutron-5c7cc4784d-fsjlq\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.654535 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.731105 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g94ln" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.805923 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.839876 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xbfz\" (UniqueName: \"kubernetes.io/projected/565cc52d-cdcf-4d59-82b5-3d71804b727a-kube-api-access-7xbfz\") pod \"565cc52d-cdcf-4d59-82b5-3d71804b727a\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.839959 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-config-data\") pod \"565cc52d-cdcf-4d59-82b5-3d71804b727a\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.840059 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-combined-ca-bundle\") pod \"565cc52d-cdcf-4d59-82b5-3d71804b727a\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.840092 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/565cc52d-cdcf-4d59-82b5-3d71804b727a-logs\") pod \"565cc52d-cdcf-4d59-82b5-3d71804b727a\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.840202 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-scripts\") pod \"565cc52d-cdcf-4d59-82b5-3d71804b727a\" (UID: \"565cc52d-cdcf-4d59-82b5-3d71804b727a\") " Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.841213 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/565cc52d-cdcf-4d59-82b5-3d71804b727a-logs" (OuterVolumeSpecName: "logs") pod "565cc52d-cdcf-4d59-82b5-3d71804b727a" (UID: "565cc52d-cdcf-4d59-82b5-3d71804b727a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.845935 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-scripts" (OuterVolumeSpecName: "scripts") pod "565cc52d-cdcf-4d59-82b5-3d71804b727a" (UID: "565cc52d-cdcf-4d59-82b5-3d71804b727a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.848156 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565cc52d-cdcf-4d59-82b5-3d71804b727a-kube-api-access-7xbfz" (OuterVolumeSpecName: "kube-api-access-7xbfz") pod "565cc52d-cdcf-4d59-82b5-3d71804b727a" (UID: "565cc52d-cdcf-4d59-82b5-3d71804b727a"). InnerVolumeSpecName "kube-api-access-7xbfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.876285 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-config-data" (OuterVolumeSpecName: "config-data") pod "565cc52d-cdcf-4d59-82b5-3d71804b727a" (UID: "565cc52d-cdcf-4d59-82b5-3d71804b727a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.881117 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "565cc52d-cdcf-4d59-82b5-3d71804b727a" (UID: "565cc52d-cdcf-4d59-82b5-3d71804b727a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.943107 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.943144 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xbfz\" (UniqueName: \"kubernetes.io/projected/565cc52d-cdcf-4d59-82b5-3d71804b727a-kube-api-access-7xbfz\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.943157 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.943166 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/565cc52d-cdcf-4d59-82b5-3d71804b727a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:18 crc kubenswrapper[5025]: I1004 10:52:18.943177 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/565cc52d-cdcf-4d59-82b5-3d71804b727a-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.153558 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"773c73f6-a2c6-46c0-9616-a8b0782ad6bf","Type":"ContainerStarted","Data":"bd28d8339d4021c8df3b131b9e1c817d7b5e0192e8664efea9de876e249e89b8"} Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.157801 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-g94ln" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.159147 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-g94ln" event={"ID":"565cc52d-cdcf-4d59-82b5-3d71804b727a","Type":"ContainerDied","Data":"850bf24e3e048dc319f2e926d50f276578ff822b45a927590b6c20f9d74f6ced"} Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.159206 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="850bf24e3e048dc319f2e926d50f276578ff822b45a927590b6c20f9d74f6ced" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.246090 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-67f4db9dc8-qprqb"] Oct 04 10:52:19 crc kubenswrapper[5025]: E1004 10:52:19.246492 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565cc52d-cdcf-4d59-82b5-3d71804b727a" containerName="placement-db-sync" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.246508 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="565cc52d-cdcf-4d59-82b5-3d71804b727a" containerName="placement-db-sync" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.246676 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="565cc52d-cdcf-4d59-82b5-3d71804b727a" containerName="placement-db-sync" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.247541 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.251047 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.251323 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.251541 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.252406 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-mc766" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.254771 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-67f4db9dc8-qprqb"] Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.256459 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.355142 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-internal-tls-certs\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.355218 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dp24\" (UniqueName: \"kubernetes.io/projected/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-kube-api-access-2dp24\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.355250 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-public-tls-certs\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.355274 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-logs\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.355630 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-combined-ca-bundle\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.355679 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-scripts\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.355723 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-config-data\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.457870 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-internal-tls-certs\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.458278 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dp24\" (UniqueName: \"kubernetes.io/projected/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-kube-api-access-2dp24\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.458306 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-public-tls-certs\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.458333 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-logs\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.458427 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-combined-ca-bundle\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.458453 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-scripts\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.458478 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-config-data\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.461398 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-logs\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.464578 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-public-tls-certs\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.464864 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-config-data\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.466743 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-combined-ca-bundle\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.466804 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-internal-tls-certs\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.465896 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-scripts\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.484640 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dp24\" (UniqueName: \"kubernetes.io/projected/3a1c727a-b835-42d9-ae4d-81f6ea3018fa-kube-api-access-2dp24\") pod \"placement-67f4db9dc8-qprqb\" (UID: \"3a1c727a-b835-42d9-ae4d-81f6ea3018fa\") " pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.594053 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.624036 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c7cc4784d-fsjlq"] Oct 04 10:52:19 crc kubenswrapper[5025]: I1004 10:52:19.637738 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-c6gbx"] Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.174135 5025 generic.go:334] "Generic (PLEG): container finished" podID="d87fc908-36a3-4364-a4e4-85bb1abb660a" containerID="27287fca9695c5c44606f27eab7953333e63519a391fa405ee3566f719e2ee99" exitCode=0 Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.175089 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" event={"ID":"d87fc908-36a3-4364-a4e4-85bb1abb660a","Type":"ContainerDied","Data":"27287fca9695c5c44606f27eab7953333e63519a391fa405ee3566f719e2ee99"} Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.175136 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" event={"ID":"d87fc908-36a3-4364-a4e4-85bb1abb660a","Type":"ContainerStarted","Data":"a6919aa94ffe7b8e8f1922936daa73b562d7e45af27b5b892226b4d385ea2ca7"} Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.178367 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"773c73f6-a2c6-46c0-9616-a8b0782ad6bf","Type":"ContainerStarted","Data":"1ae3a21b0cdb53e4bedf82ae70616497d5d1a0516a45bf2c84906ffe42903193"} Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.179825 5025 generic.go:334] "Generic (PLEG): container finished" podID="4a4861ec-2e6e-408d-9b42-5ce3248fd640" containerID="27f95935eab866d09895b9f363cc4d687040656a7e9cf672deb677ef1396860a" exitCode=0 Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.179877 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dncjm" event={"ID":"4a4861ec-2e6e-408d-9b42-5ce3248fd640","Type":"ContainerDied","Data":"27f95935eab866d09895b9f363cc4d687040656a7e9cf672deb677ef1396860a"} Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.183088 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sfm5z" event={"ID":"b5d127d5-e426-48fe-bb22-c811de5c801d","Type":"ContainerStarted","Data":"013fcc42e6f1203c89c3b5e9723bc4b707a00316525e047d9ef581503f1ca4c3"} Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.193390 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7cc4784d-fsjlq" event={"ID":"6622de76-40c3-434a-a87d-127e3d50fd4e","Type":"ContainerStarted","Data":"d2e7e42ce3b54ff6adac03225c622649ea27ad669b54ec29b2e602a8ed356ed8"} Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.193441 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7cc4784d-fsjlq" event={"ID":"6622de76-40c3-434a-a87d-127e3d50fd4e","Type":"ContainerStarted","Data":"efd69e51ce1dc3aa6d08770135fb98a0cfdb524c35023f6a01f55ea7f5f542ac"} Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.227155 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-67f4db9dc8-qprqb"] Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.228583 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-sfm5z" podStartSLOduration=2.827533734 podStartE2EDuration="31.228567833s" podCreationTimestamp="2025-10-04 10:51:49 +0000 UTC" firstStartedPulling="2025-10-04 10:51:50.556095227 +0000 UTC m=+1038.981062107" lastFinishedPulling="2025-10-04 10:52:18.957129326 +0000 UTC m=+1067.382096206" observedRunningTime="2025-10-04 10:52:20.215182072 +0000 UTC m=+1068.640148952" watchObservedRunningTime="2025-10-04 10:52:20.228567833 +0000 UTC m=+1068.653534713" Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.782706 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.783030 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.937091 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:20 crc kubenswrapper[5025]: I1004 10:52:20.937135 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.226027 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67f4db9dc8-qprqb" event={"ID":"3a1c727a-b835-42d9-ae4d-81f6ea3018fa","Type":"ContainerStarted","Data":"d77ad174d092040a6ec5ed9b8577f15287202e2c5ce4a4600885968200faae4c"} Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.226082 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67f4db9dc8-qprqb" event={"ID":"3a1c727a-b835-42d9-ae4d-81f6ea3018fa","Type":"ContainerStarted","Data":"8741c952192f481faa4a7b1ce0c0879dd32c7e873548e1e72237fe048d29f4ce"} Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.234920 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7cc4784d-fsjlq" event={"ID":"6622de76-40c3-434a-a87d-127e3d50fd4e","Type":"ContainerStarted","Data":"ff03deccd6b2c3a461ece1ebc4065b040461ecd81634ac290dc01ea1a65e5819"} Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.235310 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.240846 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c7b9c48cc-j7mmw"] Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.246851 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" event={"ID":"d87fc908-36a3-4364-a4e4-85bb1abb660a","Type":"ContainerStarted","Data":"9599e792034366698ef8820b7faae2795782557237319bc10d68bba9acec0774"} Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.246906 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.246998 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.251309 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.251480 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.256069 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"773c73f6-a2c6-46c0-9616-a8b0782ad6bf","Type":"ContainerStarted","Data":"04134829f18094ec61789f563b5ffeba54b42c2452193c675283e042fa4241a2"} Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.257871 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c7b9c48cc-j7mmw"] Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.266675 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c7cc4784d-fsjlq" podStartSLOduration=3.26665796 podStartE2EDuration="3.26665796s" podCreationTimestamp="2025-10-04 10:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:21.257591125 +0000 UTC m=+1069.682558005" watchObservedRunningTime="2025-10-04 10:52:21.26665796 +0000 UTC m=+1069.691624840" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.298954 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" podStartSLOduration=3.298937274 podStartE2EDuration="3.298937274s" podCreationTimestamp="2025-10-04 10:52:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:21.295365609 +0000 UTC m=+1069.720332499" watchObservedRunningTime="2025-10-04 10:52:21.298937274 +0000 UTC m=+1069.723904154" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.322755 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.322736889 podStartE2EDuration="4.322736889s" podCreationTimestamp="2025-10-04 10:52:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:21.321471012 +0000 UTC m=+1069.746437902" watchObservedRunningTime="2025-10-04 10:52:21.322736889 +0000 UTC m=+1069.747703769" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.415504 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-ovndb-tls-certs\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.415577 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-httpd-config\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.415660 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-internal-tls-certs\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.415694 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-config\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.415782 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-combined-ca-bundle\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.415813 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-public-tls-certs\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.415839 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsx44\" (UniqueName: \"kubernetes.io/projected/993d2c5b-1d22-4185-a265-2fe9bffc0318-kube-api-access-bsx44\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.526892 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-combined-ca-bundle\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.526939 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-public-tls-certs\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.526962 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsx44\" (UniqueName: \"kubernetes.io/projected/993d2c5b-1d22-4185-a265-2fe9bffc0318-kube-api-access-bsx44\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.527064 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-ovndb-tls-certs\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.527091 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-httpd-config\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.527140 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-internal-tls-certs\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.527155 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-config\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.535799 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-public-tls-certs\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.536401 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-combined-ca-bundle\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.539646 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-httpd-config\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.540199 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-internal-tls-certs\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.549103 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-config\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.557450 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsx44\" (UniqueName: \"kubernetes.io/projected/993d2c5b-1d22-4185-a265-2fe9bffc0318-kube-api-access-bsx44\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.563911 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/993d2c5b-1d22-4185-a265-2fe9bffc0318-ovndb-tls-certs\") pod \"neutron-c7b9c48cc-j7mmw\" (UID: \"993d2c5b-1d22-4185-a265-2fe9bffc0318\") " pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:21 crc kubenswrapper[5025]: I1004 10:52:21.571635 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:23 crc kubenswrapper[5025]: I1004 10:52:23.649046 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:23 crc kubenswrapper[5025]: I1004 10:52:23.649317 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:23 crc kubenswrapper[5025]: I1004 10:52:23.690041 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:23 crc kubenswrapper[5025]: I1004 10:52:23.697059 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:24 crc kubenswrapper[5025]: I1004 10:52:24.206204 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:52:24 crc kubenswrapper[5025]: I1004 10:52:24.279942 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:24 crc kubenswrapper[5025]: I1004 10:52:24.279992 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.299134 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dncjm" event={"ID":"4a4861ec-2e6e-408d-9b42-5ce3248fd640","Type":"ContainerDied","Data":"d8342b5433e0d1474ee87088d4056cba0fbcdbe10f7c3566f6f75fd061fd7d1b"} Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.299407 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8342b5433e0d1474ee87088d4056cba0fbcdbe10f7c3566f6f75fd061fd7d1b" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.321682 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.490389 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2frd7\" (UniqueName: \"kubernetes.io/projected/4a4861ec-2e6e-408d-9b42-5ce3248fd640-kube-api-access-2frd7\") pod \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.490449 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-fernet-keys\") pod \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.490524 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-combined-ca-bundle\") pod \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.490555 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-credential-keys\") pod \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.490609 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-config-data\") pod \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.490686 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-scripts\") pod \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\" (UID: \"4a4861ec-2e6e-408d-9b42-5ce3248fd640\") " Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.496706 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a4861ec-2e6e-408d-9b42-5ce3248fd640-kube-api-access-2frd7" (OuterVolumeSpecName: "kube-api-access-2frd7") pod "4a4861ec-2e6e-408d-9b42-5ce3248fd640" (UID: "4a4861ec-2e6e-408d-9b42-5ce3248fd640"). InnerVolumeSpecName "kube-api-access-2frd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.497118 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "4a4861ec-2e6e-408d-9b42-5ce3248fd640" (UID: "4a4861ec-2e6e-408d-9b42-5ce3248fd640"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.497154 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "4a4861ec-2e6e-408d-9b42-5ce3248fd640" (UID: "4a4861ec-2e6e-408d-9b42-5ce3248fd640"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.497927 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-scripts" (OuterVolumeSpecName: "scripts") pod "4a4861ec-2e6e-408d-9b42-5ce3248fd640" (UID: "4a4861ec-2e6e-408d-9b42-5ce3248fd640"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.520783 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-config-data" (OuterVolumeSpecName: "config-data") pod "4a4861ec-2e6e-408d-9b42-5ce3248fd640" (UID: "4a4861ec-2e6e-408d-9b42-5ce3248fd640"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.523581 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4a4861ec-2e6e-408d-9b42-5ce3248fd640" (UID: "4a4861ec-2e6e-408d-9b42-5ce3248fd640"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.593580 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.593617 5025 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.593626 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.593637 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.593648 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2frd7\" (UniqueName: \"kubernetes.io/projected/4a4861ec-2e6e-408d-9b42-5ce3248fd640-kube-api-access-2frd7\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.593658 5025 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a4861ec-2e6e-408d-9b42-5ce3248fd640-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:25 crc kubenswrapper[5025]: I1004 10:52:25.703895 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c7b9c48cc-j7mmw"] Oct 04 10:52:25 crc kubenswrapper[5025]: W1004 10:52:25.710531 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod993d2c5b_1d22_4185_a265_2fe9bffc0318.slice/crio-0c8dde3319438c3c94c9a6f26be4f61481881e20d4990ee835f5d5b4e1ba5d27 WatchSource:0}: Error finding container 0c8dde3319438c3c94c9a6f26be4f61481881e20d4990ee835f5d5b4e1ba5d27: Status 404 returned error can't find the container with id 0c8dde3319438c3c94c9a6f26be4f61481881e20d4990ee835f5d5b4e1ba5d27 Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.251090 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.255773 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.313446 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-67f4db9dc8-qprqb" event={"ID":"3a1c727a-b835-42d9-ae4d-81f6ea3018fa","Type":"ContainerStarted","Data":"59be27d89fe472754e4dc1352e1439a3a60c9e36f0d0ce98cd65db6d3088de8e"} Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.313931 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.313964 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.317245 5025 generic.go:334] "Generic (PLEG): container finished" podID="b5d127d5-e426-48fe-bb22-c811de5c801d" containerID="013fcc42e6f1203c89c3b5e9723bc4b707a00316525e047d9ef581503f1ca4c3" exitCode=0 Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.317297 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sfm5z" event={"ID":"b5d127d5-e426-48fe-bb22-c811de5c801d","Type":"ContainerDied","Data":"013fcc42e6f1203c89c3b5e9723bc4b707a00316525e047d9ef581503f1ca4c3"} Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.322845 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03dee37f-5b63-4b21-b401-f05b07d3b1af","Type":"ContainerStarted","Data":"d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230"} Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.324968 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dncjm" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.326109 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7b9c48cc-j7mmw" event={"ID":"993d2c5b-1d22-4185-a265-2fe9bffc0318","Type":"ContainerStarted","Data":"fba00fecdfebb1cda679894f66d508ffcc0ed7d94ab28c15a56fcf6282d20d5a"} Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.326158 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7b9c48cc-j7mmw" event={"ID":"993d2c5b-1d22-4185-a265-2fe9bffc0318","Type":"ContainerStarted","Data":"0c8dde3319438c3c94c9a6f26be4f61481881e20d4990ee835f5d5b4e1ba5d27"} Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.360536 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-67f4db9dc8-qprqb" podStartSLOduration=7.360510866 podStartE2EDuration="7.360510866s" podCreationTimestamp="2025-10-04 10:52:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:26.348379882 +0000 UTC m=+1074.773346762" watchObservedRunningTime="2025-10-04 10:52:26.360510866 +0000 UTC m=+1074.785477746" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.447245 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6f856fd68f-ckb8t"] Oct 04 10:52:26 crc kubenswrapper[5025]: E1004 10:52:26.447641 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a4861ec-2e6e-408d-9b42-5ce3248fd640" containerName="keystone-bootstrap" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.447654 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a4861ec-2e6e-408d-9b42-5ce3248fd640" containerName="keystone-bootstrap" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.447820 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a4861ec-2e6e-408d-9b42-5ce3248fd640" containerName="keystone-bootstrap" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.448354 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.453064 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6f856fd68f-ckb8t"] Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.455188 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.455233 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.455188 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.455350 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.455600 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9d9vx" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.455786 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.514008 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-config-data\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.514089 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-scripts\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.514134 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-credential-keys\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.514151 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-combined-ca-bundle\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.514176 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrctg\" (UniqueName: \"kubernetes.io/projected/15a00f7c-91b2-4635-b46b-8ae22b23f14e-kube-api-access-rrctg\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.514201 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-fernet-keys\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.514288 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-internal-tls-certs\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.514306 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-public-tls-certs\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.615608 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-credential-keys\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.615657 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-combined-ca-bundle\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.615691 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrctg\" (UniqueName: \"kubernetes.io/projected/15a00f7c-91b2-4635-b46b-8ae22b23f14e-kube-api-access-rrctg\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.615729 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-fernet-keys\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.615840 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-internal-tls-certs\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.615864 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-public-tls-certs\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.615899 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-config-data\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.615928 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-scripts\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.621189 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-scripts\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.624616 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-combined-ca-bundle\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.625541 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-public-tls-certs\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.627548 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-fernet-keys\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.628420 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-internal-tls-certs\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.631397 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-credential-keys\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.631503 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a00f7c-91b2-4635-b46b-8ae22b23f14e-config-data\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.641082 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrctg\" (UniqueName: \"kubernetes.io/projected/15a00f7c-91b2-4635-b46b-8ae22b23f14e-kube-api-access-rrctg\") pod \"keystone-6f856fd68f-ckb8t\" (UID: \"15a00f7c-91b2-4635-b46b-8ae22b23f14e\") " pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:26 crc kubenswrapper[5025]: I1004 10:52:26.785955 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.312893 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6f856fd68f-ckb8t"] Oct 04 10:52:27 crc kubenswrapper[5025]: W1004 10:52:27.327161 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15a00f7c_91b2_4635_b46b_8ae22b23f14e.slice/crio-727c6f2b0e3ff9dd7ab94c8f62a961202f13b18bdad8b3a83785544a712b9440 WatchSource:0}: Error finding container 727c6f2b0e3ff9dd7ab94c8f62a961202f13b18bdad8b3a83785544a712b9440: Status 404 returned error can't find the container with id 727c6f2b0e3ff9dd7ab94c8f62a961202f13b18bdad8b3a83785544a712b9440 Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.333996 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c7b9c48cc-j7mmw" event={"ID":"993d2c5b-1d22-4185-a265-2fe9bffc0318","Type":"ContainerStarted","Data":"1b5f75e98a58c052d71e02c3ef5cf30e03eb0a68e3dd9d1f38b4317a016fb8ce"} Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.334483 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.386904 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c7b9c48cc-j7mmw" podStartSLOduration=6.386880642 podStartE2EDuration="6.386880642s" podCreationTimestamp="2025-10-04 10:52:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:27.370166553 +0000 UTC m=+1075.795133423" watchObservedRunningTime="2025-10-04 10:52:27.386880642 +0000 UTC m=+1075.811847522" Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.513319 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.513425 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.549402 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.574700 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.835220 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.946255 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-combined-ca-bundle\") pod \"b5d127d5-e426-48fe-bb22-c811de5c801d\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.946459 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-db-sync-config-data\") pod \"b5d127d5-e426-48fe-bb22-c811de5c801d\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.946556 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7mnr\" (UniqueName: \"kubernetes.io/projected/b5d127d5-e426-48fe-bb22-c811de5c801d-kube-api-access-w7mnr\") pod \"b5d127d5-e426-48fe-bb22-c811de5c801d\" (UID: \"b5d127d5-e426-48fe-bb22-c811de5c801d\") " Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.954142 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b5d127d5-e426-48fe-bb22-c811de5c801d" (UID: "b5d127d5-e426-48fe-bb22-c811de5c801d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.954271 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d127d5-e426-48fe-bb22-c811de5c801d-kube-api-access-w7mnr" (OuterVolumeSpecName: "kube-api-access-w7mnr") pod "b5d127d5-e426-48fe-bb22-c811de5c801d" (UID: "b5d127d5-e426-48fe-bb22-c811de5c801d"). InnerVolumeSpecName "kube-api-access-w7mnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:27 crc kubenswrapper[5025]: I1004 10:52:27.979641 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5d127d5-e426-48fe-bb22-c811de5c801d" (UID: "b5d127d5-e426-48fe-bb22-c811de5c801d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.050328 5025 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.050363 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7mnr\" (UniqueName: \"kubernetes.io/projected/b5d127d5-e426-48fe-bb22-c811de5c801d-kube-api-access-w7mnr\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.050400 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d127d5-e426-48fe-bb22-c811de5c801d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.302430 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.348051 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6f856fd68f-ckb8t" event={"ID":"15a00f7c-91b2-4635-b46b-8ae22b23f14e","Type":"ContainerStarted","Data":"8d11008618e712d00585589e2f7d5aa87a74aabaf68c779bbefb38d52abbd232"} Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.348089 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6f856fd68f-ckb8t" event={"ID":"15a00f7c-91b2-4635-b46b-8ae22b23f14e","Type":"ContainerStarted","Data":"727c6f2b0e3ff9dd7ab94c8f62a961202f13b18bdad8b3a83785544a712b9440"} Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.348322 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.351616 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-sfm5z" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.352795 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-sfm5z" event={"ID":"b5d127d5-e426-48fe-bb22-c811de5c801d","Type":"ContainerDied","Data":"6ece5abf1e402fe3af16910eaee087f8451761283ab612edac0c69c5c6907eda"} Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.352820 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ece5abf1e402fe3af16910eaee087f8451761283ab612edac0c69c5c6907eda" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.353167 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.353405 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.404726 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6f856fd68f-ckb8t" podStartSLOduration=2.404708677 podStartE2EDuration="2.404708677s" podCreationTimestamp="2025-10-04 10:52:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:28.370427685 +0000 UTC m=+1076.795394565" watchObservedRunningTime="2025-10-04 10:52:28.404708677 +0000 UTC m=+1076.829675557" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.656136 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.659675 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt"] Oct 04 10:52:28 crc kubenswrapper[5025]: E1004 10:52:28.660059 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d127d5-e426-48fe-bb22-c811de5c801d" containerName="barbican-db-sync" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.660073 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d127d5-e426-48fe-bb22-c811de5c801d" containerName="barbican-db-sync" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.660256 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d127d5-e426-48fe-bb22-c811de5c801d" containerName="barbican-db-sync" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.661118 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.666962 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.667189 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-27v2q" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.667342 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.697548 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6db97b4dc9-fxss9"] Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.699033 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.703563 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.713417 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt"] Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.748085 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6db97b4dc9-fxss9"] Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.769153 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cfb63d9-275c-4019-8809-7c90883cab5d-logs\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.769848 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cfb63d9-275c-4019-8809-7c90883cab5d-config-data\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.769885 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cfb63d9-275c-4019-8809-7c90883cab5d-config-data-custom\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.769922 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8sk4\" (UniqueName: \"kubernetes.io/projected/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-kube-api-access-m8sk4\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.769982 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-combined-ca-bundle\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.770036 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdpsx\" (UniqueName: \"kubernetes.io/projected/0cfb63d9-275c-4019-8809-7c90883cab5d-kube-api-access-tdpsx\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.770071 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cfb63d9-275c-4019-8809-7c90883cab5d-combined-ca-bundle\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.770089 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-logs\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.770112 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-config-data-custom\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.770130 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-config-data\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.778167 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-tspdr"] Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.778490 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" podUID="510866f0-02ee-48ef-8219-fdea80015c5a" containerName="dnsmasq-dns" containerID="cri-o://3b59104bcfd14dfd14e3f04666af878bf92f5b2ec64c1efe09add8a1ddf9bfcd" gracePeriod=10 Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.798561 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-phzxk"] Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.800519 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.804977 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-phzxk"] Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.871879 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-combined-ca-bundle\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.871940 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdpsx\" (UniqueName: \"kubernetes.io/projected/0cfb63d9-275c-4019-8809-7c90883cab5d-kube-api-access-tdpsx\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.871967 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cfb63d9-275c-4019-8809-7c90883cab5d-combined-ca-bundle\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.871985 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-logs\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873535 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873565 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-config-data-custom\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873584 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-config-data\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873609 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-svc\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873639 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2stm\" (UniqueName: \"kubernetes.io/projected/46b5a923-1064-4149-b4c5-44bfc9822814-kube-api-access-v2stm\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873655 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873673 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cfb63d9-275c-4019-8809-7c90883cab5d-logs\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873701 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cfb63d9-275c-4019-8809-7c90883cab5d-config-data\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873737 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cfb63d9-275c-4019-8809-7c90883cab5d-config-data-custom\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873771 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8sk4\" (UniqueName: \"kubernetes.io/projected/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-kube-api-access-m8sk4\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873795 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-config\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.873819 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.874511 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0cfb63d9-275c-4019-8809-7c90883cab5d-logs\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.877948 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-logs\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.884254 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-config-data-custom\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.884865 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cfb63d9-275c-4019-8809-7c90883cab5d-combined-ca-bundle\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.887677 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cfb63d9-275c-4019-8809-7c90883cab5d-config-data-custom\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.888587 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cfb63d9-275c-4019-8809-7c90883cab5d-config-data\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.888704 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-combined-ca-bundle\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.897411 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdpsx\" (UniqueName: \"kubernetes.io/projected/0cfb63d9-275c-4019-8809-7c90883cab5d-kube-api-access-tdpsx\") pod \"barbican-worker-6db97b4dc9-fxss9\" (UID: \"0cfb63d9-275c-4019-8809-7c90883cab5d\") " pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.901098 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8sk4\" (UniqueName: \"kubernetes.io/projected/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-kube-api-access-m8sk4\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.927355 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5c856cb876-bgqs2"] Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.932044 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6acf4ca7-9b65-42cb-8a0d-d09b14292c6e-config-data\") pod \"barbican-keystone-listener-74bc6fc8d4-xzvxt\" (UID: \"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e\") " pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.935904 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.946092 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5c856cb876-bgqs2"] Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.968456 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.976418 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20ea1497-01a3-4d84-9bba-e7abf404c787-logs\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.976461 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.976501 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data-custom\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.976525 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-config\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.976552 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.976604 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.976638 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-svc\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.977631 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-config\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.977806 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.980425 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.980573 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-svc\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.976659 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cktkj\" (UniqueName: \"kubernetes.io/projected/20ea1497-01a3-4d84-9bba-e7abf404c787-kube-api-access-cktkj\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.987644 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2stm\" (UniqueName: \"kubernetes.io/projected/46b5a923-1064-4149-b4c5-44bfc9822814-kube-api-access-v2stm\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.987673 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.987720 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-combined-ca-bundle\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:28 crc kubenswrapper[5025]: I1004 10:52:28.988615 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:28.999727 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.032854 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2stm\" (UniqueName: \"kubernetes.io/projected/46b5a923-1064-4149-b4c5-44bfc9822814-kube-api-access-v2stm\") pod \"dnsmasq-dns-688c87cc99-phzxk\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.081596 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6db97b4dc9-fxss9" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.094314 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data-custom\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.094416 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cktkj\" (UniqueName: \"kubernetes.io/projected/20ea1497-01a3-4d84-9bba-e7abf404c787-kube-api-access-cktkj\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.094451 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-combined-ca-bundle\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.094488 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20ea1497-01a3-4d84-9bba-e7abf404c787-logs\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.094509 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.098791 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20ea1497-01a3-4d84-9bba-e7abf404c787-logs\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.105584 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-combined-ca-bundle\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.105668 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data-custom\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.107661 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.124335 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cktkj\" (UniqueName: \"kubernetes.io/projected/20ea1497-01a3-4d84-9bba-e7abf404c787-kube-api-access-cktkj\") pod \"barbican-api-5c856cb876-bgqs2\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.140250 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:29 crc kubenswrapper[5025]: E1004 10:52:29.164291 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod510866f0_02ee_48ef_8219_fdea80015c5a.slice/crio-conmon-3b59104bcfd14dfd14e3f04666af878bf92f5b2ec64c1efe09add8a1ddf9bfcd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod510866f0_02ee_48ef_8219_fdea80015c5a.slice/crio-3b59104bcfd14dfd14e3f04666af878bf92f5b2ec64c1efe09add8a1ddf9bfcd.scope\": RecentStats: unable to find data in memory cache]" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.417424 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.454201 5025 generic.go:334] "Generic (PLEG): container finished" podID="510866f0-02ee-48ef-8219-fdea80015c5a" containerID="3b59104bcfd14dfd14e3f04666af878bf92f5b2ec64c1efe09add8a1ddf9bfcd" exitCode=0 Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.454470 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" event={"ID":"510866f0-02ee-48ef-8219-fdea80015c5a","Type":"ContainerDied","Data":"3b59104bcfd14dfd14e3f04666af878bf92f5b2ec64c1efe09add8a1ddf9bfcd"} Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.566292 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.618520 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt"] Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.625544 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwtpd\" (UniqueName: \"kubernetes.io/projected/510866f0-02ee-48ef-8219-fdea80015c5a-kube-api-access-mwtpd\") pod \"510866f0-02ee-48ef-8219-fdea80015c5a\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.625618 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-svc\") pod \"510866f0-02ee-48ef-8219-fdea80015c5a\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.625687 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-sb\") pod \"510866f0-02ee-48ef-8219-fdea80015c5a\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.625820 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-swift-storage-0\") pod \"510866f0-02ee-48ef-8219-fdea80015c5a\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.625849 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-config\") pod \"510866f0-02ee-48ef-8219-fdea80015c5a\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.625869 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-nb\") pod \"510866f0-02ee-48ef-8219-fdea80015c5a\" (UID: \"510866f0-02ee-48ef-8219-fdea80015c5a\") " Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.646557 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/510866f0-02ee-48ef-8219-fdea80015c5a-kube-api-access-mwtpd" (OuterVolumeSpecName: "kube-api-access-mwtpd") pod "510866f0-02ee-48ef-8219-fdea80015c5a" (UID: "510866f0-02ee-48ef-8219-fdea80015c5a"). InnerVolumeSpecName "kube-api-access-mwtpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.679810 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "510866f0-02ee-48ef-8219-fdea80015c5a" (UID: "510866f0-02ee-48ef-8219-fdea80015c5a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.687629 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "510866f0-02ee-48ef-8219-fdea80015c5a" (UID: "510866f0-02ee-48ef-8219-fdea80015c5a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.698893 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "510866f0-02ee-48ef-8219-fdea80015c5a" (UID: "510866f0-02ee-48ef-8219-fdea80015c5a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.712316 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "510866f0-02ee-48ef-8219-fdea80015c5a" (UID: "510866f0-02ee-48ef-8219-fdea80015c5a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.724444 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-config" (OuterVolumeSpecName: "config") pod "510866f0-02ee-48ef-8219-fdea80015c5a" (UID: "510866f0-02ee-48ef-8219-fdea80015c5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.727706 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwtpd\" (UniqueName: \"kubernetes.io/projected/510866f0-02ee-48ef-8219-fdea80015c5a-kube-api-access-mwtpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.727730 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.727742 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.727749 5025 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.727757 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.727766 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/510866f0-02ee-48ef-8219-fdea80015c5a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.785312 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-phzxk"] Oct 04 10:52:29 crc kubenswrapper[5025]: I1004 10:52:29.901530 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6db97b4dc9-fxss9"] Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.076223 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5c856cb876-bgqs2"] Oct 04 10:52:30 crc kubenswrapper[5025]: W1004 10:52:30.097382 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ea1497_01a3_4d84_9bba_e7abf404c787.slice/crio-9ade0ebbc11e91b24f82234254042433133bbdafa63970231cc97b8c0d7ea4b9 WatchSource:0}: Error finding container 9ade0ebbc11e91b24f82234254042433133bbdafa63970231cc97b8c0d7ea4b9: Status 404 returned error can't find the container with id 9ade0ebbc11e91b24f82234254042433133bbdafa63970231cc97b8c0d7ea4b9 Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.470794 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c856cb876-bgqs2" event={"ID":"20ea1497-01a3-4d84-9bba-e7abf404c787","Type":"ContainerStarted","Data":"42c2292ecadfbc91dc0e9c87d4f8e32fcb6859f8c5de5d74f458cae466267bf1"} Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.472915 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c856cb876-bgqs2" event={"ID":"20ea1497-01a3-4d84-9bba-e7abf404c787","Type":"ContainerStarted","Data":"9ade0ebbc11e91b24f82234254042433133bbdafa63970231cc97b8c0d7ea4b9"} Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.482618 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" event={"ID":"510866f0-02ee-48ef-8219-fdea80015c5a","Type":"ContainerDied","Data":"05112581d44898b9a0f956529afc46d25f85b2c6f4f62a85b4c443073d7fb567"} Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.482680 5025 scope.go:117] "RemoveContainer" containerID="3b59104bcfd14dfd14e3f04666af878bf92f5b2ec64c1efe09add8a1ddf9bfcd" Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.482871 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-tspdr" Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.499819 5025 generic.go:334] "Generic (PLEG): container finished" podID="46b5a923-1064-4149-b4c5-44bfc9822814" containerID="077579dc999df40c9545c9a495aa1280f22c86d5c13d55a4978ca71c6f397e6e" exitCode=0 Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.499870 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" event={"ID":"46b5a923-1064-4149-b4c5-44bfc9822814","Type":"ContainerDied","Data":"077579dc999df40c9545c9a495aa1280f22c86d5c13d55a4978ca71c6f397e6e"} Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.499894 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" event={"ID":"46b5a923-1064-4149-b4c5-44bfc9822814","Type":"ContainerStarted","Data":"a11daa0ac17bc9a0250470723ff68198d6e9522022874b898c5db841f7a1ff7e"} Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.502286 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" event={"ID":"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e","Type":"ContainerStarted","Data":"daa162b7372e1edb8505f551d16b54e4b1a61af045cf7066031406e2c0e2398b"} Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.515331 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6db97b4dc9-fxss9" event={"ID":"0cfb63d9-275c-4019-8809-7c90883cab5d","Type":"ContainerStarted","Data":"4c58506dd264598648d5bdaf1c195b490a47a36253d20a2344c5b4ef2b5e39bc"} Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.544082 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-tspdr"] Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.550774 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-tspdr"] Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.581347 5025 scope.go:117] "RemoveContainer" containerID="1c155077832c655cf8b156dff2fa8d96206cca3c55adff9fef5052513b07d4e4" Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.770090 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.770141 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.787171 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5787d4b6b6-fc88x" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 10:52:30 crc kubenswrapper[5025]: I1004 10:52:30.938366 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6559967db4-zfm7d" podUID="53944185-08b0-4249-8662-9fdce7cb24cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.532167 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c856cb876-bgqs2" event={"ID":"20ea1497-01a3-4d84-9bba-e7abf404c787","Type":"ContainerStarted","Data":"4146b8c8b42f82b144405ec2d0a537654909fd161ae073e045f001b05fa05cd0"} Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.533074 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.533098 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.541221 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xtspb" event={"ID":"694e381e-49b8-49f1-a5ae-7aa3b72d8fee","Type":"ContainerStarted","Data":"2b783084ad2ac2c644297be5333b2f46bdda1ea8bdaa25217af12d8ae5abc93a"} Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.554525 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5c856cb876-bgqs2" podStartSLOduration=3.5545095890000002 podStartE2EDuration="3.554509589s" podCreationTimestamp="2025-10-04 10:52:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:31.552653565 +0000 UTC m=+1079.977620445" watchObservedRunningTime="2025-10-04 10:52:31.554509589 +0000 UTC m=+1079.979476469" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.563914 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" event={"ID":"46b5a923-1064-4149-b4c5-44bfc9822814","Type":"ContainerStarted","Data":"989d54a75b1e0e03a97c0d08c7395ec92c1633dfc1bb1936b112d2cf2f6a9aee"} Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.564139 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.582792 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-xtspb" podStartSLOduration=3.25132111 podStartE2EDuration="42.582774705s" podCreationTimestamp="2025-10-04 10:51:49 +0000 UTC" firstStartedPulling="2025-10-04 10:51:50.726162998 +0000 UTC m=+1039.151129878" lastFinishedPulling="2025-10-04 10:52:30.057616593 +0000 UTC m=+1078.482583473" observedRunningTime="2025-10-04 10:52:31.573935967 +0000 UTC m=+1079.998902847" watchObservedRunningTime="2025-10-04 10:52:31.582774705 +0000 UTC m=+1080.007741575" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.593866 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" podStartSLOduration=3.593851279 podStartE2EDuration="3.593851279s" podCreationTimestamp="2025-10-04 10:52:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:31.591242283 +0000 UTC m=+1080.016209163" watchObservedRunningTime="2025-10-04 10:52:31.593851279 +0000 UTC m=+1080.018818159" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.955444 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6bfdcdbc7d-swl5f"] Oct 04 10:52:31 crc kubenswrapper[5025]: E1004 10:52:31.955804 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="510866f0-02ee-48ef-8219-fdea80015c5a" containerName="dnsmasq-dns" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.955819 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="510866f0-02ee-48ef-8219-fdea80015c5a" containerName="dnsmasq-dns" Oct 04 10:52:31 crc kubenswrapper[5025]: E1004 10:52:31.955831 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="510866f0-02ee-48ef-8219-fdea80015c5a" containerName="init" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.955837 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="510866f0-02ee-48ef-8219-fdea80015c5a" containerName="init" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.956464 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="510866f0-02ee-48ef-8219-fdea80015c5a" containerName="dnsmasq-dns" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.957374 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.959584 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.963347 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.970620 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6bfdcdbc7d-swl5f"] Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.987957 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-public-tls-certs\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.988025 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvhgj\" (UniqueName: \"kubernetes.io/projected/fd52b707-ec59-4653-974d-9b0963fee564-kube-api-access-nvhgj\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.988047 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd52b707-ec59-4653-974d-9b0963fee564-logs\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.988077 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-internal-tls-certs\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.988144 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-config-data\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.988223 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-combined-ca-bundle\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:31 crc kubenswrapper[5025]: I1004 10:52:31.988255 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-config-data-custom\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.091034 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-combined-ca-bundle\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.091112 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-config-data-custom\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.091143 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-public-tls-certs\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.091180 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvhgj\" (UniqueName: \"kubernetes.io/projected/fd52b707-ec59-4653-974d-9b0963fee564-kube-api-access-nvhgj\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.091199 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd52b707-ec59-4653-974d-9b0963fee564-logs\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.091232 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-internal-tls-certs\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.091313 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-config-data\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.093709 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd52b707-ec59-4653-974d-9b0963fee564-logs\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.097602 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-public-tls-certs\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.098271 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-internal-tls-certs\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.098467 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-config-data\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.098544 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-config-data-custom\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.099034 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd52b707-ec59-4653-974d-9b0963fee564-combined-ca-bundle\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.112217 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvhgj\" (UniqueName: \"kubernetes.io/projected/fd52b707-ec59-4653-974d-9b0963fee564-kube-api-access-nvhgj\") pod \"barbican-api-6bfdcdbc7d-swl5f\" (UID: \"fd52b707-ec59-4653-974d-9b0963fee564\") " pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.275752 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.475250 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="510866f0-02ee-48ef-8219-fdea80015c5a" path="/var/lib/kubelet/pods/510866f0-02ee-48ef-8219-fdea80015c5a/volumes" Oct 04 10:52:32 crc kubenswrapper[5025]: I1004 10:52:32.921130 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6bfdcdbc7d-swl5f"] Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.602538 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" event={"ID":"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e","Type":"ContainerStarted","Data":"1bad950fac38f9b540ae2c1f75b694bab56c661ad0b56772a8b7f96f7300eaa4"} Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.602854 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" event={"ID":"6acf4ca7-9b65-42cb-8a0d-d09b14292c6e","Type":"ContainerStarted","Data":"ad08c6d09259d1ae870e151aa76256233c2b37e887dbd62379ef56cc574f4880"} Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.610430 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bfdcdbc7d-swl5f" event={"ID":"fd52b707-ec59-4653-974d-9b0963fee564","Type":"ContainerStarted","Data":"b0b529dd92133efbc0274ba74325ef43276db96166153da657d56f0ae68a4ecd"} Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.610678 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bfdcdbc7d-swl5f" event={"ID":"fd52b707-ec59-4653-974d-9b0963fee564","Type":"ContainerStarted","Data":"5f10e577cb77befabdf4d2ab9d98e622606b17ca368692e2101777c3578fae77"} Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.610690 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6bfdcdbc7d-swl5f" event={"ID":"fd52b707-ec59-4653-974d-9b0963fee564","Type":"ContainerStarted","Data":"e33d20ab9dc9bde8b7abeccdd54fc09f593f4ec7f123985136b0c9fce42f499c"} Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.610823 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.610926 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.619650 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6db97b4dc9-fxss9" event={"ID":"0cfb63d9-275c-4019-8809-7c90883cab5d","Type":"ContainerStarted","Data":"9eb00024f52e1f8ca64264aeeefa84491f460756553dfd370af32f099c58758e"} Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.619687 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6db97b4dc9-fxss9" event={"ID":"0cfb63d9-275c-4019-8809-7c90883cab5d","Type":"ContainerStarted","Data":"6054334b23e82c42dd2cef04e5f063704f04537935da7c76cc761ff7f1223023"} Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.623526 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-74bc6fc8d4-xzvxt" podStartSLOduration=2.848407582 podStartE2EDuration="5.623512444s" podCreationTimestamp="2025-10-04 10:52:28 +0000 UTC" firstStartedPulling="2025-10-04 10:52:29.629238793 +0000 UTC m=+1078.054205673" lastFinishedPulling="2025-10-04 10:52:32.404343655 +0000 UTC m=+1080.829310535" observedRunningTime="2025-10-04 10:52:33.616004195 +0000 UTC m=+1082.040971075" watchObservedRunningTime="2025-10-04 10:52:33.623512444 +0000 UTC m=+1082.048479324" Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.647908 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6bfdcdbc7d-swl5f" podStartSLOduration=2.647887977 podStartE2EDuration="2.647887977s" podCreationTimestamp="2025-10-04 10:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:33.637998108 +0000 UTC m=+1082.062964988" watchObservedRunningTime="2025-10-04 10:52:33.647887977 +0000 UTC m=+1082.072854857" Oct 04 10:52:33 crc kubenswrapper[5025]: I1004 10:52:33.668933 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6db97b4dc9-fxss9" podStartSLOduration=3.174016289 podStartE2EDuration="5.668914221s" podCreationTimestamp="2025-10-04 10:52:28 +0000 UTC" firstStartedPulling="2025-10-04 10:52:29.915421807 +0000 UTC m=+1078.340388687" lastFinishedPulling="2025-10-04 10:52:32.410319739 +0000 UTC m=+1080.835286619" observedRunningTime="2025-10-04 10:52:33.660550837 +0000 UTC m=+1082.085517717" watchObservedRunningTime="2025-10-04 10:52:33.668914221 +0000 UTC m=+1082.093881101" Oct 04 10:52:36 crc kubenswrapper[5025]: I1004 10:52:36.662064 5025 generic.go:334] "Generic (PLEG): container finished" podID="694e381e-49b8-49f1-a5ae-7aa3b72d8fee" containerID="2b783084ad2ac2c644297be5333b2f46bdda1ea8bdaa25217af12d8ae5abc93a" exitCode=0 Oct 04 10:52:36 crc kubenswrapper[5025]: I1004 10:52:36.662152 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xtspb" event={"ID":"694e381e-49b8-49f1-a5ae-7aa3b72d8fee","Type":"ContainerDied","Data":"2b783084ad2ac2c644297be5333b2f46bdda1ea8bdaa25217af12d8ae5abc93a"} Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.352173 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xtspb" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.563291 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-config-data\") pod \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.563502 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68hzg\" (UniqueName: \"kubernetes.io/projected/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-kube-api-access-68hzg\") pod \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.563540 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-db-sync-config-data\") pod \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.563581 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-scripts\") pod \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.563611 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-etc-machine-id\") pod \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.563635 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-combined-ca-bundle\") pod \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\" (UID: \"694e381e-49b8-49f1-a5ae-7aa3b72d8fee\") " Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.564251 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "694e381e-49b8-49f1-a5ae-7aa3b72d8fee" (UID: "694e381e-49b8-49f1-a5ae-7aa3b72d8fee"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.564373 5025 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.568142 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-scripts" (OuterVolumeSpecName: "scripts") pod "694e381e-49b8-49f1-a5ae-7aa3b72d8fee" (UID: "694e381e-49b8-49f1-a5ae-7aa3b72d8fee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.568564 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "694e381e-49b8-49f1-a5ae-7aa3b72d8fee" (UID: "694e381e-49b8-49f1-a5ae-7aa3b72d8fee"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.569469 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-kube-api-access-68hzg" (OuterVolumeSpecName: "kube-api-access-68hzg") pod "694e381e-49b8-49f1-a5ae-7aa3b72d8fee" (UID: "694e381e-49b8-49f1-a5ae-7aa3b72d8fee"). InnerVolumeSpecName "kube-api-access-68hzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.605109 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "694e381e-49b8-49f1-a5ae-7aa3b72d8fee" (UID: "694e381e-49b8-49f1-a5ae-7aa3b72d8fee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.621554 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-config-data" (OuterVolumeSpecName: "config-data") pod "694e381e-49b8-49f1-a5ae-7aa3b72d8fee" (UID: "694e381e-49b8-49f1-a5ae-7aa3b72d8fee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:38 crc kubenswrapper[5025]: E1004 10:52:38.633781 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.667586 5025 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.667851 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.667953 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.668052 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.668151 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68hzg\" (UniqueName: \"kubernetes.io/projected/694e381e-49b8-49f1-a5ae-7aa3b72d8fee-kube-api-access-68hzg\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.681782 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03dee37f-5b63-4b21-b401-f05b07d3b1af","Type":"ContainerStarted","Data":"dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062"} Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.682091 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="ceilometer-notification-agent" containerID="cri-o://872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33" gracePeriod=30 Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.682188 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="proxy-httpd" containerID="cri-o://dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062" gracePeriod=30 Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.682230 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="sg-core" containerID="cri-o://d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230" gracePeriod=30 Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.682322 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.687662 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-xtspb" event={"ID":"694e381e-49b8-49f1-a5ae-7aa3b72d8fee","Type":"ContainerDied","Data":"c03095558b2e2a07df1cf4fb37ac72b9357c6c9d45470610d2cf0bbf0ab11242"} Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.687816 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c03095558b2e2a07df1cf4fb37ac72b9357c6c9d45470610d2cf0bbf0ab11242" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.687742 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-xtspb" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.974849 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 10:52:38 crc kubenswrapper[5025]: E1004 10:52:38.975236 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="694e381e-49b8-49f1-a5ae-7aa3b72d8fee" containerName="cinder-db-sync" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.975256 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="694e381e-49b8-49f1-a5ae-7aa3b72d8fee" containerName="cinder-db-sync" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.975423 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="694e381e-49b8-49f1-a5ae-7aa3b72d8fee" containerName="cinder-db-sync" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.983482 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.994587 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.994753 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.994789 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 04 10:52:38 crc kubenswrapper[5025]: I1004 10:52:38.995172 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-jnpdb" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.002755 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.082234 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-phzxk"] Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.082466 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" podUID="46b5a923-1064-4149-b4c5-44bfc9822814" containerName="dnsmasq-dns" containerID="cri-o://989d54a75b1e0e03a97c0d08c7395ec92c1633dfc1bb1936b112d2cf2f6a9aee" gracePeriod=10 Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.082510 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.082828 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06f8ff3c-7082-4c52-bd79-6bf11327631e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.082990 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.083080 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-scripts\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.083216 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.083315 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7swd\" (UniqueName: \"kubernetes.io/projected/06f8ff3c-7082-4c52-bd79-6bf11327631e-kube-api-access-x7swd\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.084276 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.133153 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-89zdw"] Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.134770 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.161110 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-89zdw"] Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.185580 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7swd\" (UniqueName: \"kubernetes.io/projected/06f8ff3c-7082-4c52-bd79-6bf11327631e-kube-api-access-x7swd\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.185814 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.185916 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.186023 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.186110 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.186190 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.186265 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06f8ff3c-7082-4c52-bd79-6bf11327631e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.186346 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-config\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.186415 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.186487 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-scripts\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.186572 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7q59\" (UniqueName: \"kubernetes.io/projected/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-kube-api-access-s7q59\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.186656 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.189153 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06f8ff3c-7082-4c52-bd79-6bf11327631e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.193692 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.194712 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-scripts\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.197236 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.203027 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.208811 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7swd\" (UniqueName: \"kubernetes.io/projected/06f8ff3c-7082-4c52-bd79-6bf11327631e-kube-api-access-x7swd\") pod \"cinder-scheduler-0\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.219857 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.222292 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.228652 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.231799 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.292961 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.293050 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.293512 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.293570 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.293613 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-config\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.293661 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7q59\" (UniqueName: \"kubernetes.io/projected/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-kube-api-access-s7q59\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.293899 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.295633 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.296275 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.306714 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-config\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.306934 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.308472 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7q59\" (UniqueName: \"kubernetes.io/projected/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-kube-api-access-s7q59\") pod \"dnsmasq-dns-6bb4fc677f-89zdw\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.308874 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.395002 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.395064 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.395103 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data-custom\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.395119 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-logs\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.395180 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4582v\" (UniqueName: \"kubernetes.io/projected/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-kube-api-access-4582v\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.395262 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.395316 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-scripts\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.475228 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.497110 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4582v\" (UniqueName: \"kubernetes.io/projected/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-kube-api-access-4582v\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.497192 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.497216 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-scripts\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.497261 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.497278 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.497306 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data-custom\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.497322 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-logs\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.497753 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-logs\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.498069 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.511883 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-scripts\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.513158 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.513573 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.524316 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data-custom\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.530623 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4582v\" (UniqueName: \"kubernetes.io/projected/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-kube-api-access-4582v\") pod \"cinder-api-0\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.710471 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.720467 5025 generic.go:334] "Generic (PLEG): container finished" podID="46b5a923-1064-4149-b4c5-44bfc9822814" containerID="989d54a75b1e0e03a97c0d08c7395ec92c1633dfc1bb1936b112d2cf2f6a9aee" exitCode=0 Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.720556 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" event={"ID":"46b5a923-1064-4149-b4c5-44bfc9822814","Type":"ContainerDied","Data":"989d54a75b1e0e03a97c0d08c7395ec92c1633dfc1bb1936b112d2cf2f6a9aee"} Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.780295 5025 generic.go:334] "Generic (PLEG): container finished" podID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerID="d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230" exitCode=2 Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.780338 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03dee37f-5b63-4b21-b401-f05b07d3b1af","Type":"ContainerDied","Data":"d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230"} Oct 04 10:52:39 crc kubenswrapper[5025]: I1004 10:52:39.849530 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.021727 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-89zdw"] Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.160165 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.333570 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-sb\") pod \"46b5a923-1064-4149-b4c5-44bfc9822814\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.333951 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-svc\") pod \"46b5a923-1064-4149-b4c5-44bfc9822814\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.336954 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-nb\") pod \"46b5a923-1064-4149-b4c5-44bfc9822814\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.337055 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-config\") pod \"46b5a923-1064-4149-b4c5-44bfc9822814\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.337157 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2stm\" (UniqueName: \"kubernetes.io/projected/46b5a923-1064-4149-b4c5-44bfc9822814-kube-api-access-v2stm\") pod \"46b5a923-1064-4149-b4c5-44bfc9822814\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.337317 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-swift-storage-0\") pod \"46b5a923-1064-4149-b4c5-44bfc9822814\" (UID: \"46b5a923-1064-4149-b4c5-44bfc9822814\") " Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.355797 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46b5a923-1064-4149-b4c5-44bfc9822814-kube-api-access-v2stm" (OuterVolumeSpecName: "kube-api-access-v2stm") pod "46b5a923-1064-4149-b4c5-44bfc9822814" (UID: "46b5a923-1064-4149-b4c5-44bfc9822814"). InnerVolumeSpecName "kube-api-access-v2stm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.389691 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "46b5a923-1064-4149-b4c5-44bfc9822814" (UID: "46b5a923-1064-4149-b4c5-44bfc9822814"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.402683 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "46b5a923-1064-4149-b4c5-44bfc9822814" (UID: "46b5a923-1064-4149-b4c5-44bfc9822814"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.423937 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "46b5a923-1064-4149-b4c5-44bfc9822814" (UID: "46b5a923-1064-4149-b4c5-44bfc9822814"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.432321 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-config" (OuterVolumeSpecName: "config") pod "46b5a923-1064-4149-b4c5-44bfc9822814" (UID: "46b5a923-1064-4149-b4c5-44bfc9822814"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.448801 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.448831 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.448841 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2stm\" (UniqueName: \"kubernetes.io/projected/46b5a923-1064-4149-b4c5-44bfc9822814-kube-api-access-v2stm\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.448850 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.448858 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.457565 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "46b5a923-1064-4149-b4c5-44bfc9822814" (UID: "46b5a923-1064-4149-b4c5-44bfc9822814"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.520558 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.550737 5025 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/46b5a923-1064-4149-b4c5-44bfc9822814-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.783243 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5787d4b6b6-fc88x" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.793783 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06f8ff3c-7082-4c52-bd79-6bf11327631e","Type":"ContainerStarted","Data":"b828565a66e51bf8d0f63b10d6b3c7f3949e6c965c3bfef69976ec9bada66876"} Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.804713 5025 generic.go:334] "Generic (PLEG): container finished" podID="a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" containerID="0c7c4e3c6c52fdb30cbf761a3525873752d10706c6a678374babc0582477ad78" exitCode=0 Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.804843 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" event={"ID":"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf","Type":"ContainerDied","Data":"0c7c4e3c6c52fdb30cbf761a3525873752d10706c6a678374babc0582477ad78"} Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.805305 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" event={"ID":"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf","Type":"ContainerStarted","Data":"87980b017fba990245e4a878da2ccd87fdc784a9b137d63f5e404593c6179313"} Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.815890 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" event={"ID":"46b5a923-1064-4149-b4c5-44bfc9822814","Type":"ContainerDied","Data":"a11daa0ac17bc9a0250470723ff68198d6e9522022874b898c5db841f7a1ff7e"} Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.816215 5025 scope.go:117] "RemoveContainer" containerID="989d54a75b1e0e03a97c0d08c7395ec92c1633dfc1bb1936b112d2cf2f6a9aee" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.815983 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-phzxk" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.847250 5025 generic.go:334] "Generic (PLEG): container finished" podID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerID="872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33" exitCode=0 Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.847368 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03dee37f-5b63-4b21-b401-f05b07d3b1af","Type":"ContainerDied","Data":"872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33"} Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.850294 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea","Type":"ContainerStarted","Data":"b3e0d4c8e97160e4cb0239e1fd1fc54c08b68b1392ff496234a189cd88d8f643"} Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.883278 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-phzxk"] Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.896134 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-phzxk"] Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.897094 5025 scope.go:117] "RemoveContainer" containerID="077579dc999df40c9545c9a495aa1280f22c86d5c13d55a4978ca71c6f397e6e" Oct 04 10:52:40 crc kubenswrapper[5025]: I1004 10:52:40.938855 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6559967db4-zfm7d" podUID="53944185-08b0-4249-8662-9fdce7cb24cf" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 04 10:52:41 crc kubenswrapper[5025]: I1004 10:52:41.252547 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 10:52:41 crc kubenswrapper[5025]: I1004 10:52:41.459206 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:41 crc kubenswrapper[5025]: I1004 10:52:41.492650 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:41 crc kubenswrapper[5025]: I1004 10:52:41.865219 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06f8ff3c-7082-4c52-bd79-6bf11327631e","Type":"ContainerStarted","Data":"ad6459fe128541eede78cd87fd4860229b37b095ac9aef89222cf6c720ecf3ac"} Oct 04 10:52:41 crc kubenswrapper[5025]: I1004 10:52:41.897632 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" event={"ID":"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf","Type":"ContainerStarted","Data":"3b78cd90ae0947943209e117223a1b25701dc2b8bd997f4eee6275518e3db983"} Oct 04 10:52:41 crc kubenswrapper[5025]: I1004 10:52:41.897713 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:41 crc kubenswrapper[5025]: I1004 10:52:41.906327 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea","Type":"ContainerStarted","Data":"3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f"} Oct 04 10:52:42 crc kubenswrapper[5025]: I1004 10:52:42.420427 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46b5a923-1064-4149-b4c5-44bfc9822814" path="/var/lib/kubelet/pods/46b5a923-1064-4149-b4c5-44bfc9822814/volumes" Oct 04 10:52:42 crc kubenswrapper[5025]: I1004 10:52:42.439873 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" podStartSLOduration=3.439855489 podStartE2EDuration="3.439855489s" podCreationTimestamp="2025-10-04 10:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:41.919163692 +0000 UTC m=+1090.344130572" watchObservedRunningTime="2025-10-04 10:52:42.439855489 +0000 UTC m=+1090.864822369" Oct 04 10:52:42 crc kubenswrapper[5025]: I1004 10:52:42.917099 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea","Type":"ContainerStarted","Data":"ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14"} Oct 04 10:52:42 crc kubenswrapper[5025]: I1004 10:52:42.917444 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" containerName="cinder-api-log" containerID="cri-o://3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f" gracePeriod=30 Oct 04 10:52:42 crc kubenswrapper[5025]: I1004 10:52:42.917790 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 10:52:42 crc kubenswrapper[5025]: I1004 10:52:42.917994 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" containerName="cinder-api" containerID="cri-o://ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14" gracePeriod=30 Oct 04 10:52:42 crc kubenswrapper[5025]: I1004 10:52:42.921073 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06f8ff3c-7082-4c52-bd79-6bf11327631e","Type":"ContainerStarted","Data":"ddf6a6970b20cf0149f80ba8c621f3bf930705b92594d9ee02682dfe4981d994"} Oct 04 10:52:42 crc kubenswrapper[5025]: I1004 10:52:42.940425 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.940407086 podStartE2EDuration="3.940407086s" podCreationTimestamp="2025-10-04 10:52:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:42.933043501 +0000 UTC m=+1091.358010381" watchObservedRunningTime="2025-10-04 10:52:42.940407086 +0000 UTC m=+1091.365373966" Oct 04 10:52:42 crc kubenswrapper[5025]: I1004 10:52:42.958738 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.23050703 podStartE2EDuration="4.958723382s" podCreationTimestamp="2025-10-04 10:52:38 +0000 UTC" firstStartedPulling="2025-10-04 10:52:39.890499224 +0000 UTC m=+1088.315466104" lastFinishedPulling="2025-10-04 10:52:40.618715576 +0000 UTC m=+1089.043682456" observedRunningTime="2025-10-04 10:52:42.953143589 +0000 UTC m=+1091.378110469" watchObservedRunningTime="2025-10-04 10:52:42.958723382 +0000 UTC m=+1091.383690262" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.557938 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.714078 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4582v\" (UniqueName: \"kubernetes.io/projected/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-kube-api-access-4582v\") pod \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.714154 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-etc-machine-id\") pod \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.714184 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-logs\") pod \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.714219 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-combined-ca-bundle\") pod \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.714271 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-scripts\") pod \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.714306 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data-custom\") pod \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.714288 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" (UID: "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.714366 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data\") pod \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\" (UID: \"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea\") " Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.714794 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-logs" (OuterVolumeSpecName: "logs") pod "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" (UID: "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.715552 5025 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.715582 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.720286 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-kube-api-access-4582v" (OuterVolumeSpecName: "kube-api-access-4582v") pod "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" (UID: "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea"). InnerVolumeSpecName "kube-api-access-4582v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.721240 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-scripts" (OuterVolumeSpecName: "scripts") pod "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" (UID: "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.736568 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" (UID: "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.769989 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" (UID: "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.787518 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.805712 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data" (OuterVolumeSpecName: "config-data") pod "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" (UID: "e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.820624 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.820659 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.820668 5025 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.820676 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.820685 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4582v\" (UniqueName: \"kubernetes.io/projected/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea-kube-api-access-4582v\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.930406 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6bfdcdbc7d-swl5f" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.944292 5025 generic.go:334] "Generic (PLEG): container finished" podID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" containerID="ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14" exitCode=0 Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.944322 5025 generic.go:334] "Generic (PLEG): container finished" podID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" containerID="3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f" exitCode=143 Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.945109 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.945214 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea","Type":"ContainerDied","Data":"ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14"} Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.945262 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea","Type":"ContainerDied","Data":"3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f"} Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.945285 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea","Type":"ContainerDied","Data":"b3e0d4c8e97160e4cb0239e1fd1fc54c08b68b1392ff496234a189cd88d8f643"} Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.945326 5025 scope.go:117] "RemoveContainer" containerID="ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.987585 5025 scope.go:117] "RemoveContainer" containerID="3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f" Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.991288 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5c856cb876-bgqs2"] Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.991541 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5c856cb876-bgqs2" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerName="barbican-api-log" containerID="cri-o://42c2292ecadfbc91dc0e9c87d4f8e32fcb6859f8c5de5d74f458cae466267bf1" gracePeriod=30 Oct 04 10:52:43 crc kubenswrapper[5025]: I1004 10:52:43.991703 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5c856cb876-bgqs2" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerName="barbican-api" containerID="cri-o://4146b8c8b42f82b144405ec2d0a537654909fd161ae073e045f001b05fa05cd0" gracePeriod=30 Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.008838 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.057231 5025 scope.go:117] "RemoveContainer" containerID="ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.057355 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.057393 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 04 10:52:44 crc kubenswrapper[5025]: E1004 10:52:44.057881 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" containerName="cinder-api-log" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.057904 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" containerName="cinder-api-log" Oct 04 10:52:44 crc kubenswrapper[5025]: E1004 10:52:44.057932 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b5a923-1064-4149-b4c5-44bfc9822814" containerName="init" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.057941 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b5a923-1064-4149-b4c5-44bfc9822814" containerName="init" Oct 04 10:52:44 crc kubenswrapper[5025]: E1004 10:52:44.057960 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b5a923-1064-4149-b4c5-44bfc9822814" containerName="dnsmasq-dns" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.057969 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b5a923-1064-4149-b4c5-44bfc9822814" containerName="dnsmasq-dns" Oct 04 10:52:44 crc kubenswrapper[5025]: E1004 10:52:44.057984 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" containerName="cinder-api" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.057992 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" containerName="cinder-api" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.058215 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="46b5a923-1064-4149-b4c5-44bfc9822814" containerName="dnsmasq-dns" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.058256 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" containerName="cinder-api-log" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.058274 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" containerName="cinder-api" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.059502 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: E1004 10:52:44.063157 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14\": container with ID starting with ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14 not found: ID does not exist" containerID="ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.063231 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14"} err="failed to get container status \"ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14\": rpc error: code = NotFound desc = could not find container \"ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14\": container with ID starting with ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14 not found: ID does not exist" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.063257 5025 scope.go:117] "RemoveContainer" containerID="3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f" Oct 04 10:52:44 crc kubenswrapper[5025]: E1004 10:52:44.064722 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f\": container with ID starting with 3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f not found: ID does not exist" containerID="3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.064762 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f"} err="failed to get container status \"3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f\": rpc error: code = NotFound desc = could not find container \"3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f\": container with ID starting with 3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f not found: ID does not exist" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.064793 5025 scope.go:117] "RemoveContainer" containerID="ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.065315 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.065362 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.066219 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.066237 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14"} err="failed to get container status \"ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14\": rpc error: code = NotFound desc = could not find container \"ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14\": container with ID starting with ad01dadb0084567dc373889013a366b47bba6bed0dfee7e0f3a8d0ea4252bd14 not found: ID does not exist" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.066265 5025 scope.go:117] "RemoveContainer" containerID="3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.066745 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f"} err="failed to get container status \"3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f\": rpc error: code = NotFound desc = could not find container \"3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f\": container with ID starting with 3e23c07823ddfe99a7fa9b68666b209f5ecafae3d82e52341a4750ac8896605f not found: ID does not exist" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.072245 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.233755 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-scripts\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.233906 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxn2w\" (UniqueName: \"kubernetes.io/projected/3f3cd68e-1753-4b60-94ba-4c33d2a95034-kube-api-access-wxn2w\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.233943 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-config-data-custom\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.233980 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.233997 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f3cd68e-1753-4b60-94ba-4c33d2a95034-logs\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.234031 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.234254 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f3cd68e-1753-4b60-94ba-4c33d2a95034-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.234359 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-config-data\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.234400 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.309637 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.335904 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.335949 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f3cd68e-1753-4b60-94ba-4c33d2a95034-logs\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.335982 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.336048 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f3cd68e-1753-4b60-94ba-4c33d2a95034-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.336076 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-config-data\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.336097 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.336136 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-scripts\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.336190 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxn2w\" (UniqueName: \"kubernetes.io/projected/3f3cd68e-1753-4b60-94ba-4c33d2a95034-kube-api-access-wxn2w\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.336209 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-config-data-custom\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.336818 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f3cd68e-1753-4b60-94ba-4c33d2a95034-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.337109 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f3cd68e-1753-4b60-94ba-4c33d2a95034-logs\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.342972 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.343078 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.343107 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.343967 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-config-data\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.344173 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-config-data-custom\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.352189 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f3cd68e-1753-4b60-94ba-4c33d2a95034-scripts\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.357686 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxn2w\" (UniqueName: \"kubernetes.io/projected/3f3cd68e-1753-4b60-94ba-4c33d2a95034-kube-api-access-wxn2w\") pod \"cinder-api-0\" (UID: \"3f3cd68e-1753-4b60-94ba-4c33d2a95034\") " pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.425335 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea" path="/var/lib/kubelet/pods/e0a13b2e-37c0-4cc0-90b2-2a7a652c24ea/volumes" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.461186 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.933751 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.954936 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f3cd68e-1753-4b60-94ba-4c33d2a95034","Type":"ContainerStarted","Data":"69fe9a1d0fd64fbe62871c90cff1684b2499e73b77c5fa0a86256408ae88e84b"} Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.958791 5025 generic.go:334] "Generic (PLEG): container finished" podID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerID="42c2292ecadfbc91dc0e9c87d4f8e32fcb6859f8c5de5d74f458cae466267bf1" exitCode=143 Oct 04 10:52:44 crc kubenswrapper[5025]: I1004 10:52:44.958914 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c856cb876-bgqs2" event={"ID":"20ea1497-01a3-4d84-9bba-e7abf404c787","Type":"ContainerDied","Data":"42c2292ecadfbc91dc0e9c87d4f8e32fcb6859f8c5de5d74f458cae466267bf1"} Oct 04 10:52:45 crc kubenswrapper[5025]: I1004 10:52:45.969333 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f3cd68e-1753-4b60-94ba-4c33d2a95034","Type":"ContainerStarted","Data":"4ba3f7b67de8110534521acdc07ed8ebf4acadb42c045fffa2ae0e9921e70ac8"} Oct 04 10:52:46 crc kubenswrapper[5025]: I1004 10:52:46.985517 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3f3cd68e-1753-4b60-94ba-4c33d2a95034","Type":"ContainerStarted","Data":"3924f830fbe679b59082eb0f48bdaa41377cea535a2092f14fbd6b38b2cb6476"} Oct 04 10:52:46 crc kubenswrapper[5025]: I1004 10:52:46.985970 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.020850 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.020820455 podStartE2EDuration="3.020820455s" podCreationTimestamp="2025-10-04 10:52:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:47.004115817 +0000 UTC m=+1095.429082737" watchObservedRunningTime="2025-10-04 10:52:47.020820455 +0000 UTC m=+1095.445787375" Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.447288 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5c856cb876-bgqs2" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:57864->10.217.0.159:9311: read: connection reset by peer" Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.447398 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5c856cb876-bgqs2" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:57866->10.217.0.159:9311: read: connection reset by peer" Oct 04 10:52:47 crc kubenswrapper[5025]: W1004 10:52:47.527482 5025 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0a13b2e_37c0_4cc0_90b2_2a7a652c24ea.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0a13b2e_37c0_4cc0_90b2_2a7a652c24ea.slice: no such file or directory Oct 04 10:52:47 crc kubenswrapper[5025]: E1004 10:52:47.700856 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod694e381e_49b8_49f1_a5ae_7aa3b72d8fee.slice/crio-conmon-2b783084ad2ac2c644297be5333b2f46bdda1ea8bdaa25217af12d8ae5abc93a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod694e381e_49b8_49f1_a5ae_7aa3b72d8fee.slice/crio-c03095558b2e2a07df1cf4fb37ac72b9357c6c9d45470610d2cf0bbf0ab11242\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03dee37f_5b63_4b21_b401_f05b07d3b1af.slice/crio-conmon-872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46b5a923_1064_4149_b4c5_44bfc9822814.slice/crio-a11daa0ac17bc9a0250470723ff68198d6e9522022874b898c5db841f7a1ff7e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ea1497_01a3_4d84_9bba_e7abf404c787.slice/crio-4146b8c8b42f82b144405ec2d0a537654909fd161ae073e045f001b05fa05cd0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a30dd6c_f419_44e2_94e2_8eca192123ae.slice/crio-conmon-d69338dcf54457fcdb1b6f100e8efeb9df6797e7ae70f2ffd70bbd09b4194103.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03dee37f_5b63_4b21_b401_f05b07d3b1af.slice/crio-conmon-d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ea1497_01a3_4d84_9bba_e7abf404c787.slice/crio-conmon-4146b8c8b42f82b144405ec2d0a537654909fd161ae073e045f001b05fa05cd0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ea1497_01a3_4d84_9bba_e7abf404c787.slice/crio-42c2292ecadfbc91dc0e9c87d4f8e32fcb6859f8c5de5d74f458cae466267bf1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a30dd6c_f419_44e2_94e2_8eca192123ae.slice/crio-conmon-0a9d774c273867c86c28003916c2dd1cd5282a66d98716a6a4b6bc2026604acd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a30dd6c_f419_44e2_94e2_8eca192123ae.slice/crio-d69338dcf54457fcdb1b6f100e8efeb9df6797e7ae70f2ffd70bbd09b4194103.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03dee37f_5b63_4b21_b401_f05b07d3b1af.slice/crio-d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod694e381e_49b8_49f1_a5ae_7aa3b72d8fee.slice/crio-2b783084ad2ac2c644297be5333b2f46bdda1ea8bdaa25217af12d8ae5abc93a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46b5a923_1064_4149_b4c5_44bfc9822814.slice/crio-989d54a75b1e0e03a97c0d08c7395ec92c1633dfc1bb1936b112d2cf2f6a9aee.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20ea1497_01a3_4d84_9bba_e7abf404c787.slice/crio-conmon-42c2292ecadfbc91dc0e9c87d4f8e32fcb6859f8c5de5d74f458cae466267bf1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03dee37f_5b63_4b21_b401_f05b07d3b1af.slice/crio-872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod694e381e_49b8_49f1_a5ae_7aa3b72d8fee.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46b5a923_1064_4149_b4c5_44bfc9822814.slice/crio-conmon-989d54a75b1e0e03a97c0d08c7395ec92c1633dfc1bb1936b112d2cf2f6a9aee.scope\": RecentStats: unable to find data in memory cache]" Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.993874 5025 generic.go:334] "Generic (PLEG): container finished" podID="8a30dd6c-f419-44e2-94e2-8eca192123ae" containerID="d69338dcf54457fcdb1b6f100e8efeb9df6797e7ae70f2ffd70bbd09b4194103" exitCode=137 Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.994137 5025 generic.go:334] "Generic (PLEG): container finished" podID="8a30dd6c-f419-44e2-94e2-8eca192123ae" containerID="0a9d774c273867c86c28003916c2dd1cd5282a66d98716a6a4b6bc2026604acd" exitCode=137 Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.993942 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5964556bf-mbrwb" event={"ID":"8a30dd6c-f419-44e2-94e2-8eca192123ae","Type":"ContainerDied","Data":"d69338dcf54457fcdb1b6f100e8efeb9df6797e7ae70f2ffd70bbd09b4194103"} Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.994195 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5964556bf-mbrwb" event={"ID":"8a30dd6c-f419-44e2-94e2-8eca192123ae","Type":"ContainerDied","Data":"0a9d774c273867c86c28003916c2dd1cd5282a66d98716a6a4b6bc2026604acd"} Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.995842 5025 generic.go:334] "Generic (PLEG): container finished" podID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerID="4146b8c8b42f82b144405ec2d0a537654909fd161ae073e045f001b05fa05cd0" exitCode=0 Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.995880 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c856cb876-bgqs2" event={"ID":"20ea1497-01a3-4d84-9bba-e7abf404c787","Type":"ContainerDied","Data":"4146b8c8b42f82b144405ec2d0a537654909fd161ae073e045f001b05fa05cd0"} Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.995957 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c856cb876-bgqs2" event={"ID":"20ea1497-01a3-4d84-9bba-e7abf404c787","Type":"ContainerDied","Data":"9ade0ebbc11e91b24f82234254042433133bbdafa63970231cc97b8c0d7ea4b9"} Oct 04 10:52:47 crc kubenswrapper[5025]: I1004 10:52:47.995972 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ade0ebbc11e91b24f82234254042433133bbdafa63970231cc97b8c0d7ea4b9" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.059166 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.071636 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.148992 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-combined-ca-bundle\") pod \"20ea1497-01a3-4d84-9bba-e7abf404c787\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.149128 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8a30dd6c-f419-44e2-94e2-8eca192123ae-horizon-secret-key\") pod \"8a30dd6c-f419-44e2-94e2-8eca192123ae\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.149598 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-config-data\") pod \"8a30dd6c-f419-44e2-94e2-8eca192123ae\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.149678 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data\") pod \"20ea1497-01a3-4d84-9bba-e7abf404c787\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.149712 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20ea1497-01a3-4d84-9bba-e7abf404c787-logs\") pod \"20ea1497-01a3-4d84-9bba-e7abf404c787\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.149753 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qz9kd\" (UniqueName: \"kubernetes.io/projected/8a30dd6c-f419-44e2-94e2-8eca192123ae-kube-api-access-qz9kd\") pod \"8a30dd6c-f419-44e2-94e2-8eca192123ae\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.149798 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cktkj\" (UniqueName: \"kubernetes.io/projected/20ea1497-01a3-4d84-9bba-e7abf404c787-kube-api-access-cktkj\") pod \"20ea1497-01a3-4d84-9bba-e7abf404c787\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.150066 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a30dd6c-f419-44e2-94e2-8eca192123ae-logs\") pod \"8a30dd6c-f419-44e2-94e2-8eca192123ae\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.150130 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-scripts\") pod \"8a30dd6c-f419-44e2-94e2-8eca192123ae\" (UID: \"8a30dd6c-f419-44e2-94e2-8eca192123ae\") " Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.150164 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data-custom\") pod \"20ea1497-01a3-4d84-9bba-e7abf404c787\" (UID: \"20ea1497-01a3-4d84-9bba-e7abf404c787\") " Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.150426 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ea1497-01a3-4d84-9bba-e7abf404c787-logs" (OuterVolumeSpecName: "logs") pod "20ea1497-01a3-4d84-9bba-e7abf404c787" (UID: "20ea1497-01a3-4d84-9bba-e7abf404c787"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.150868 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a30dd6c-f419-44e2-94e2-8eca192123ae-logs" (OuterVolumeSpecName: "logs") pod "8a30dd6c-f419-44e2-94e2-8eca192123ae" (UID: "8a30dd6c-f419-44e2-94e2-8eca192123ae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.151357 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a30dd6c-f419-44e2-94e2-8eca192123ae-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.151375 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20ea1497-01a3-4d84-9bba-e7abf404c787-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.155967 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a30dd6c-f419-44e2-94e2-8eca192123ae-kube-api-access-qz9kd" (OuterVolumeSpecName: "kube-api-access-qz9kd") pod "8a30dd6c-f419-44e2-94e2-8eca192123ae" (UID: "8a30dd6c-f419-44e2-94e2-8eca192123ae"). InnerVolumeSpecName "kube-api-access-qz9kd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.155970 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "20ea1497-01a3-4d84-9bba-e7abf404c787" (UID: "20ea1497-01a3-4d84-9bba-e7abf404c787"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.157156 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ea1497-01a3-4d84-9bba-e7abf404c787-kube-api-access-cktkj" (OuterVolumeSpecName: "kube-api-access-cktkj") pod "20ea1497-01a3-4d84-9bba-e7abf404c787" (UID: "20ea1497-01a3-4d84-9bba-e7abf404c787"). InnerVolumeSpecName "kube-api-access-cktkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.158161 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a30dd6c-f419-44e2-94e2-8eca192123ae-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8a30dd6c-f419-44e2-94e2-8eca192123ae" (UID: "8a30dd6c-f419-44e2-94e2-8eca192123ae"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.177733 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-config-data" (OuterVolumeSpecName: "config-data") pod "8a30dd6c-f419-44e2-94e2-8eca192123ae" (UID: "8a30dd6c-f419-44e2-94e2-8eca192123ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.185711 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-scripts" (OuterVolumeSpecName: "scripts") pod "8a30dd6c-f419-44e2-94e2-8eca192123ae" (UID: "8a30dd6c-f419-44e2-94e2-8eca192123ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.185816 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20ea1497-01a3-4d84-9bba-e7abf404c787" (UID: "20ea1497-01a3-4d84-9bba-e7abf404c787"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.211371 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data" (OuterVolumeSpecName: "config-data") pod "20ea1497-01a3-4d84-9bba-e7abf404c787" (UID: "20ea1497-01a3-4d84-9bba-e7abf404c787"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.252749 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.252778 5025 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8a30dd6c-f419-44e2-94e2-8eca192123ae-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.252791 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.252801 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.252811 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qz9kd\" (UniqueName: \"kubernetes.io/projected/8a30dd6c-f419-44e2-94e2-8eca192123ae-kube-api-access-qz9kd\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.252824 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cktkj\" (UniqueName: \"kubernetes.io/projected/20ea1497-01a3-4d84-9bba-e7abf404c787-kube-api-access-cktkj\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.253457 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a30dd6c-f419-44e2-94e2-8eca192123ae-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.253588 5025 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/20ea1497-01a3-4d84-9bba-e7abf404c787-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:48 crc kubenswrapper[5025]: I1004 10:52:48.814463 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.013648 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c856cb876-bgqs2" Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.014149 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5964556bf-mbrwb" event={"ID":"8a30dd6c-f419-44e2-94e2-8eca192123ae","Type":"ContainerDied","Data":"cc8dc546c2918adef5c44bb5948375bbb362c06046a5ecc529bd37e6197ee9cf"} Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.014198 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5964556bf-mbrwb" Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.014203 5025 scope.go:117] "RemoveContainer" containerID="d69338dcf54457fcdb1b6f100e8efeb9df6797e7ae70f2ffd70bbd09b4194103" Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.048860 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5964556bf-mbrwb"] Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.056647 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5964556bf-mbrwb"] Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.072122 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5c856cb876-bgqs2"] Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.081217 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5c856cb876-bgqs2"] Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.230056 5025 scope.go:117] "RemoveContainer" containerID="0a9d774c273867c86c28003916c2dd1cd5282a66d98716a6a4b6bc2026604acd" Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.476270 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.551634 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-c6gbx"] Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.551912 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" podUID="d87fc908-36a3-4364-a4e4-85bb1abb660a" containerName="dnsmasq-dns" containerID="cri-o://9599e792034366698ef8820b7faae2795782557237319bc10d68bba9acec0774" gracePeriod=10 Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.565763 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.620362 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 10:52:49 crc kubenswrapper[5025]: I1004 10:52:49.733160 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.023907 5025 generic.go:334] "Generic (PLEG): container finished" podID="d87fc908-36a3-4364-a4e4-85bb1abb660a" containerID="9599e792034366698ef8820b7faae2795782557237319bc10d68bba9acec0774" exitCode=0 Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.024001 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" event={"ID":"d87fc908-36a3-4364-a4e4-85bb1abb660a","Type":"ContainerDied","Data":"9599e792034366698ef8820b7faae2795782557237319bc10d68bba9acec0774"} Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.024081 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" event={"ID":"d87fc908-36a3-4364-a4e4-85bb1abb660a","Type":"ContainerDied","Data":"a6919aa94ffe7b8e8f1922936daa73b562d7e45af27b5b892226b4d385ea2ca7"} Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.024103 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6919aa94ffe7b8e8f1922936daa73b562d7e45af27b5b892226b4d385ea2ca7" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.024153 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="06f8ff3c-7082-4c52-bd79-6bf11327631e" containerName="cinder-scheduler" containerID="cri-o://ad6459fe128541eede78cd87fd4860229b37b095ac9aef89222cf6c720ecf3ac" gracePeriod=30 Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.024247 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="06f8ff3c-7082-4c52-bd79-6bf11327631e" containerName="probe" containerID="cri-o://ddf6a6970b20cf0149f80ba8c621f3bf930705b92594d9ee02682dfe4981d994" gracePeriod=30 Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.059586 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.095432 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-svc\") pod \"d87fc908-36a3-4364-a4e4-85bb1abb660a\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.095545 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-nb\") pod \"d87fc908-36a3-4364-a4e4-85bb1abb660a\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.095648 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cncjg\" (UniqueName: \"kubernetes.io/projected/d87fc908-36a3-4364-a4e4-85bb1abb660a-kube-api-access-cncjg\") pod \"d87fc908-36a3-4364-a4e4-85bb1abb660a\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.095693 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-config\") pod \"d87fc908-36a3-4364-a4e4-85bb1abb660a\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.095718 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-sb\") pod \"d87fc908-36a3-4364-a4e4-85bb1abb660a\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.095744 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-swift-storage-0\") pod \"d87fc908-36a3-4364-a4e4-85bb1abb660a\" (UID: \"d87fc908-36a3-4364-a4e4-85bb1abb660a\") " Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.114580 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d87fc908-36a3-4364-a4e4-85bb1abb660a-kube-api-access-cncjg" (OuterVolumeSpecName: "kube-api-access-cncjg") pod "d87fc908-36a3-4364-a4e4-85bb1abb660a" (UID: "d87fc908-36a3-4364-a4e4-85bb1abb660a"). InnerVolumeSpecName "kube-api-access-cncjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.163993 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d87fc908-36a3-4364-a4e4-85bb1abb660a" (UID: "d87fc908-36a3-4364-a4e4-85bb1abb660a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.170135 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d87fc908-36a3-4364-a4e4-85bb1abb660a" (UID: "d87fc908-36a3-4364-a4e4-85bb1abb660a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.175399 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d87fc908-36a3-4364-a4e4-85bb1abb660a" (UID: "d87fc908-36a3-4364-a4e4-85bb1abb660a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.190277 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d87fc908-36a3-4364-a4e4-85bb1abb660a" (UID: "d87fc908-36a3-4364-a4e4-85bb1abb660a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.195465 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-config" (OuterVolumeSpecName: "config") pod "d87fc908-36a3-4364-a4e4-85bb1abb660a" (UID: "d87fc908-36a3-4364-a4e4-85bb1abb660a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.197807 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.197829 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cncjg\" (UniqueName: \"kubernetes.io/projected/d87fc908-36a3-4364-a4e4-85bb1abb660a-kube-api-access-cncjg\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.197840 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.197848 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.197856 5025 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.197863 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d87fc908-36a3-4364-a4e4-85bb1abb660a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.426567 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" path="/var/lib/kubelet/pods/20ea1497-01a3-4d84-9bba-e7abf404c787/volumes" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.427200 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a30dd6c-f419-44e2-94e2-8eca192123ae" path="/var/lib/kubelet/pods/8a30dd6c-f419-44e2-94e2-8eca192123ae/volumes" Oct 04 10:52:50 crc kubenswrapper[5025]: I1004 10:52:50.557060 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-67f4db9dc8-qprqb" Oct 04 10:52:51 crc kubenswrapper[5025]: I1004 10:52:51.035478 5025 generic.go:334] "Generic (PLEG): container finished" podID="06f8ff3c-7082-4c52-bd79-6bf11327631e" containerID="ddf6a6970b20cf0149f80ba8c621f3bf930705b92594d9ee02682dfe4981d994" exitCode=0 Oct 04 10:52:51 crc kubenswrapper[5025]: I1004 10:52:51.035671 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06f8ff3c-7082-4c52-bd79-6bf11327631e","Type":"ContainerDied","Data":"ddf6a6970b20cf0149f80ba8c621f3bf930705b92594d9ee02682dfe4981d994"} Oct 04 10:52:51 crc kubenswrapper[5025]: I1004 10:52:51.035818 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-c6gbx" Oct 04 10:52:51 crc kubenswrapper[5025]: I1004 10:52:51.082778 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-c6gbx"] Oct 04 10:52:51 crc kubenswrapper[5025]: I1004 10:52:51.093533 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-c6gbx"] Oct 04 10:52:51 crc kubenswrapper[5025]: I1004 10:52:51.588732 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-c7b9c48cc-j7mmw" Oct 04 10:52:51 crc kubenswrapper[5025]: I1004 10:52:51.670005 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c7cc4784d-fsjlq"] Oct 04 10:52:51 crc kubenswrapper[5025]: I1004 10:52:51.670486 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c7cc4784d-fsjlq" podUID="6622de76-40c3-434a-a87d-127e3d50fd4e" containerName="neutron-httpd" containerID="cri-o://ff03deccd6b2c3a461ece1ebc4065b040461ecd81634ac290dc01ea1a65e5819" gracePeriod=30 Oct 04 10:52:51 crc kubenswrapper[5025]: I1004 10:52:51.670839 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c7cc4784d-fsjlq" podUID="6622de76-40c3-434a-a87d-127e3d50fd4e" containerName="neutron-api" containerID="cri-o://d2e7e42ce3b54ff6adac03225c622649ea27ad669b54ec29b2e602a8ed356ed8" gracePeriod=30 Oct 04 10:52:52 crc kubenswrapper[5025]: I1004 10:52:52.045454 5025 generic.go:334] "Generic (PLEG): container finished" podID="6622de76-40c3-434a-a87d-127e3d50fd4e" containerID="ff03deccd6b2c3a461ece1ebc4065b040461ecd81634ac290dc01ea1a65e5819" exitCode=0 Oct 04 10:52:52 crc kubenswrapper[5025]: I1004 10:52:52.045513 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7cc4784d-fsjlq" event={"ID":"6622de76-40c3-434a-a87d-127e3d50fd4e","Type":"ContainerDied","Data":"ff03deccd6b2c3a461ece1ebc4065b040461ecd81634ac290dc01ea1a65e5819"} Oct 04 10:52:52 crc kubenswrapper[5025]: I1004 10:52:52.423421 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d87fc908-36a3-4364-a4e4-85bb1abb660a" path="/var/lib/kubelet/pods/d87fc908-36a3-4364-a4e4-85bb1abb660a/volumes" Oct 04 10:52:52 crc kubenswrapper[5025]: I1004 10:52:52.707761 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:52 crc kubenswrapper[5025]: I1004 10:52:52.919825 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.060941 5025 generic.go:334] "Generic (PLEG): container finished" podID="06f8ff3c-7082-4c52-bd79-6bf11327631e" containerID="ad6459fe128541eede78cd87fd4860229b37b095ac9aef89222cf6c720ecf3ac" exitCode=0 Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.061024 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06f8ff3c-7082-4c52-bd79-6bf11327631e","Type":"ContainerDied","Data":"ad6459fe128541eede78cd87fd4860229b37b095ac9aef89222cf6c720ecf3ac"} Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.207438 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.254316 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06f8ff3c-7082-4c52-bd79-6bf11327631e-etc-machine-id\") pod \"06f8ff3c-7082-4c52-bd79-6bf11327631e\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.254390 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-scripts\") pod \"06f8ff3c-7082-4c52-bd79-6bf11327631e\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.254551 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06f8ff3c-7082-4c52-bd79-6bf11327631e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "06f8ff3c-7082-4c52-bd79-6bf11327631e" (UID: "06f8ff3c-7082-4c52-bd79-6bf11327631e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.254560 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data\") pod \"06f8ff3c-7082-4c52-bd79-6bf11327631e\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.254628 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7swd\" (UniqueName: \"kubernetes.io/projected/06f8ff3c-7082-4c52-bd79-6bf11327631e-kube-api-access-x7swd\") pod \"06f8ff3c-7082-4c52-bd79-6bf11327631e\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.254652 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-combined-ca-bundle\") pod \"06f8ff3c-7082-4c52-bd79-6bf11327631e\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.254715 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data-custom\") pod \"06f8ff3c-7082-4c52-bd79-6bf11327631e\" (UID: \"06f8ff3c-7082-4c52-bd79-6bf11327631e\") " Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.255449 5025 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/06f8ff3c-7082-4c52-bd79-6bf11327631e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.274040 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-scripts" (OuterVolumeSpecName: "scripts") pod "06f8ff3c-7082-4c52-bd79-6bf11327631e" (UID: "06f8ff3c-7082-4c52-bd79-6bf11327631e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.279914 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f8ff3c-7082-4c52-bd79-6bf11327631e-kube-api-access-x7swd" (OuterVolumeSpecName: "kube-api-access-x7swd") pod "06f8ff3c-7082-4c52-bd79-6bf11327631e" (UID: "06f8ff3c-7082-4c52-bd79-6bf11327631e"). InnerVolumeSpecName "kube-api-access-x7swd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.284801 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "06f8ff3c-7082-4c52-bd79-6bf11327631e" (UID: "06f8ff3c-7082-4c52-bd79-6bf11327631e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.318301 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06f8ff3c-7082-4c52-bd79-6bf11327631e" (UID: "06f8ff3c-7082-4c52-bd79-6bf11327631e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.357072 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7swd\" (UniqueName: \"kubernetes.io/projected/06f8ff3c-7082-4c52-bd79-6bf11327631e-kube-api-access-x7swd\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.357094 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.357103 5025 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.357112 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.388403 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data" (OuterVolumeSpecName: "config-data") pod "06f8ff3c-7082-4c52-bd79-6bf11327631e" (UID: "06f8ff3c-7082-4c52-bd79-6bf11327631e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:53 crc kubenswrapper[5025]: I1004 10:52:53.458642 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06f8ff3c-7082-4c52-bd79-6bf11327631e-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.072673 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"06f8ff3c-7082-4c52-bd79-6bf11327631e","Type":"ContainerDied","Data":"b828565a66e51bf8d0f63b10d6b3c7f3949e6c965c3bfef69976ec9bada66876"} Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.073063 5025 scope.go:117] "RemoveContainer" containerID="ddf6a6970b20cf0149f80ba8c621f3bf930705b92594d9ee02682dfe4981d994" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.072910 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.101164 5025 scope.go:117] "RemoveContainer" containerID="ad6459fe128541eede78cd87fd4860229b37b095ac9aef89222cf6c720ecf3ac" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.200079 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.206875 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.240793 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 10:52:54 crc kubenswrapper[5025]: E1004 10:52:54.241369 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerName="barbican-api" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241385 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerName="barbican-api" Oct 04 10:52:54 crc kubenswrapper[5025]: E1004 10:52:54.241399 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87fc908-36a3-4364-a4e4-85bb1abb660a" containerName="init" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241405 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87fc908-36a3-4364-a4e4-85bb1abb660a" containerName="init" Oct 04 10:52:54 crc kubenswrapper[5025]: E1004 10:52:54.241414 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a30dd6c-f419-44e2-94e2-8eca192123ae" containerName="horizon" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241420 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a30dd6c-f419-44e2-94e2-8eca192123ae" containerName="horizon" Oct 04 10:52:54 crc kubenswrapper[5025]: E1004 10:52:54.241433 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f8ff3c-7082-4c52-bd79-6bf11327631e" containerName="cinder-scheduler" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241439 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f8ff3c-7082-4c52-bd79-6bf11327631e" containerName="cinder-scheduler" Oct 04 10:52:54 crc kubenswrapper[5025]: E1004 10:52:54.241459 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f8ff3c-7082-4c52-bd79-6bf11327631e" containerName="probe" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241465 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f8ff3c-7082-4c52-bd79-6bf11327631e" containerName="probe" Oct 04 10:52:54 crc kubenswrapper[5025]: E1004 10:52:54.241479 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerName="barbican-api-log" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241485 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerName="barbican-api-log" Oct 04 10:52:54 crc kubenswrapper[5025]: E1004 10:52:54.241495 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87fc908-36a3-4364-a4e4-85bb1abb660a" containerName="dnsmasq-dns" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241501 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87fc908-36a3-4364-a4e4-85bb1abb660a" containerName="dnsmasq-dns" Oct 04 10:52:54 crc kubenswrapper[5025]: E1004 10:52:54.241513 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a30dd6c-f419-44e2-94e2-8eca192123ae" containerName="horizon-log" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241519 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a30dd6c-f419-44e2-94e2-8eca192123ae" containerName="horizon-log" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241672 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a30dd6c-f419-44e2-94e2-8eca192123ae" containerName="horizon-log" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241689 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerName="barbican-api-log" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241695 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="20ea1497-01a3-4d84-9bba-e7abf404c787" containerName="barbican-api" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241706 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a30dd6c-f419-44e2-94e2-8eca192123ae" containerName="horizon" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241713 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f8ff3c-7082-4c52-bd79-6bf11327631e" containerName="probe" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241722 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="d87fc908-36a3-4364-a4e4-85bb1abb660a" containerName="dnsmasq-dns" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.241733 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f8ff3c-7082-4c52-bd79-6bf11327631e" containerName="cinder-scheduler" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.242608 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.246112 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.270458 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.273456 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-scripts\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.273538 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.273610 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6da062c0-b483-4388-8759-84f44afb073b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.273651 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.273667 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-config-data\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.273684 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7kfl\" (UniqueName: \"kubernetes.io/projected/6da062c0-b483-4388-8759-84f44afb073b-kube-api-access-p7kfl\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.375351 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.375448 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6da062c0-b483-4388-8759-84f44afb073b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.375484 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.375506 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-config-data\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.375586 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6da062c0-b483-4388-8759-84f44afb073b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.376241 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7kfl\" (UniqueName: \"kubernetes.io/projected/6da062c0-b483-4388-8759-84f44afb073b-kube-api-access-p7kfl\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.376357 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-scripts\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.380946 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.380953 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.381249 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-scripts\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.381294 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6da062c0-b483-4388-8759-84f44afb073b-config-data\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.400479 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7kfl\" (UniqueName: \"kubernetes.io/projected/6da062c0-b483-4388-8759-84f44afb073b-kube-api-access-p7kfl\") pod \"cinder-scheduler-0\" (UID: \"6da062c0-b483-4388-8759-84f44afb073b\") " pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.426256 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f8ff3c-7082-4c52-bd79-6bf11327631e" path="/var/lib/kubelet/pods/06f8ff3c-7082-4c52-bd79-6bf11327631e/volumes" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.558737 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.559210 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.839075 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6559967db4-zfm7d" Oct 04 10:52:54 crc kubenswrapper[5025]: I1004 10:52:54.899583 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5787d4b6b6-fc88x"] Oct 04 10:52:55 crc kubenswrapper[5025]: I1004 10:52:55.039839 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 04 10:52:55 crc kubenswrapper[5025]: I1004 10:52:55.089128 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6da062c0-b483-4388-8759-84f44afb073b","Type":"ContainerStarted","Data":"8cfb99c5863fe625d199a332b18ce898fc64f2cdbd02150fe78557367d1d61fe"} Oct 04 10:52:55 crc kubenswrapper[5025]: I1004 10:52:55.089190 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5787d4b6b6-fc88x" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon-log" containerID="cri-o://16fdc70e3b1d7f1e5bbf6f4cd085a28452869621c0a13d1016341a4f70694f64" gracePeriod=30 Oct 04 10:52:55 crc kubenswrapper[5025]: I1004 10:52:55.089243 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5787d4b6b6-fc88x" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon" containerID="cri-o://64a255b2f77c316007bca29fbea4431bb259ddb5e8fe92d31a3ac65b6b471726" gracePeriod=30 Oct 04 10:52:56 crc kubenswrapper[5025]: I1004 10:52:56.100786 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6da062c0-b483-4388-8759-84f44afb073b","Type":"ContainerStarted","Data":"a92b7504836eb55319449250460b5df031666864226da5bcbc476e919e47c656"} Oct 04 10:52:56 crc kubenswrapper[5025]: I1004 10:52:56.303366 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.114171 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6da062c0-b483-4388-8759-84f44afb073b","Type":"ContainerStarted","Data":"f2c420c576001018a1c2864397d82999bc3187ec1a7708b359a35a3dd0e1e4b1"} Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.117334 5025 generic.go:334] "Generic (PLEG): container finished" podID="6622de76-40c3-434a-a87d-127e3d50fd4e" containerID="d2e7e42ce3b54ff6adac03225c622649ea27ad669b54ec29b2e602a8ed356ed8" exitCode=0 Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.117386 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7cc4784d-fsjlq" event={"ID":"6622de76-40c3-434a-a87d-127e3d50fd4e","Type":"ContainerDied","Data":"d2e7e42ce3b54ff6adac03225c622649ea27ad669b54ec29b2e602a8ed356ed8"} Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.143571 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.1435524089999998 podStartE2EDuration="3.143552409s" podCreationTimestamp="2025-10-04 10:52:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:52:57.132282358 +0000 UTC m=+1105.557249258" watchObservedRunningTime="2025-10-04 10:52:57.143552409 +0000 UTC m=+1105.568519289" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.579522 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.633037 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfv7n\" (UniqueName: \"kubernetes.io/projected/6622de76-40c3-434a-a87d-127e3d50fd4e-kube-api-access-zfv7n\") pod \"6622de76-40c3-434a-a87d-127e3d50fd4e\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.633108 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-config\") pod \"6622de76-40c3-434a-a87d-127e3d50fd4e\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.633133 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-httpd-config\") pod \"6622de76-40c3-434a-a87d-127e3d50fd4e\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.633162 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-ovndb-tls-certs\") pod \"6622de76-40c3-434a-a87d-127e3d50fd4e\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.633247 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-combined-ca-bundle\") pod \"6622de76-40c3-434a-a87d-127e3d50fd4e\" (UID: \"6622de76-40c3-434a-a87d-127e3d50fd4e\") " Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.638642 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6622de76-40c3-434a-a87d-127e3d50fd4e" (UID: "6622de76-40c3-434a-a87d-127e3d50fd4e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.663778 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6622de76-40c3-434a-a87d-127e3d50fd4e-kube-api-access-zfv7n" (OuterVolumeSpecName: "kube-api-access-zfv7n") pod "6622de76-40c3-434a-a87d-127e3d50fd4e" (UID: "6622de76-40c3-434a-a87d-127e3d50fd4e"). InnerVolumeSpecName "kube-api-access-zfv7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.683862 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6622de76-40c3-434a-a87d-127e3d50fd4e" (UID: "6622de76-40c3-434a-a87d-127e3d50fd4e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.693203 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-config" (OuterVolumeSpecName: "config") pod "6622de76-40c3-434a-a87d-127e3d50fd4e" (UID: "6622de76-40c3-434a-a87d-127e3d50fd4e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.721622 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6622de76-40c3-434a-a87d-127e3d50fd4e" (UID: "6622de76-40c3-434a-a87d-127e3d50fd4e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.735125 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.735149 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfv7n\" (UniqueName: \"kubernetes.io/projected/6622de76-40c3-434a-a87d-127e3d50fd4e-kube-api-access-zfv7n\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.735160 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.735169 5025 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:57 crc kubenswrapper[5025]: I1004 10:52:57.735176 5025 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6622de76-40c3-434a-a87d-127e3d50fd4e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:52:58 crc kubenswrapper[5025]: I1004 10:52:58.147844 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c7cc4784d-fsjlq" Oct 04 10:52:58 crc kubenswrapper[5025]: I1004 10:52:58.154153 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c7cc4784d-fsjlq" event={"ID":"6622de76-40c3-434a-a87d-127e3d50fd4e","Type":"ContainerDied","Data":"efd69e51ce1dc3aa6d08770135fb98a0cfdb524c35023f6a01f55ea7f5f542ac"} Oct 04 10:52:58 crc kubenswrapper[5025]: I1004 10:52:58.154213 5025 scope.go:117] "RemoveContainer" containerID="ff03deccd6b2c3a461ece1ebc4065b040461ecd81634ac290dc01ea1a65e5819" Oct 04 10:52:58 crc kubenswrapper[5025]: I1004 10:52:58.207679 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c7cc4784d-fsjlq"] Oct 04 10:52:58 crc kubenswrapper[5025]: I1004 10:52:58.214097 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5c7cc4784d-fsjlq"] Oct 04 10:52:58 crc kubenswrapper[5025]: I1004 10:52:58.215921 5025 scope.go:117] "RemoveContainer" containerID="d2e7e42ce3b54ff6adac03225c622649ea27ad669b54ec29b2e602a8ed356ed8" Oct 04 10:52:58 crc kubenswrapper[5025]: I1004 10:52:58.423862 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6622de76-40c3-434a-a87d-127e3d50fd4e" path="/var/lib/kubelet/pods/6622de76-40c3-434a-a87d-127e3d50fd4e/volumes" Oct 04 10:52:58 crc kubenswrapper[5025]: I1004 10:52:58.440145 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6f856fd68f-ckb8t" Oct 04 10:52:59 crc kubenswrapper[5025]: I1004 10:52:59.158912 5025 generic.go:334] "Generic (PLEG): container finished" podID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerID="64a255b2f77c316007bca29fbea4431bb259ddb5e8fe92d31a3ac65b6b471726" exitCode=0 Oct 04 10:52:59 crc kubenswrapper[5025]: I1004 10:52:59.158961 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5787d4b6b6-fc88x" event={"ID":"1a4337f7-6dc7-4f98-8633-a61b61010693","Type":"ContainerDied","Data":"64a255b2f77c316007bca29fbea4431bb259ddb5e8fe92d31a3ac65b6b471726"} Oct 04 10:52:59 crc kubenswrapper[5025]: I1004 10:52:59.559475 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 04 10:53:00 crc kubenswrapper[5025]: I1004 10:53:00.512857 5025 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod510866f0-02ee-48ef-8219-fdea80015c5a"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod510866f0-02ee-48ef-8219-fdea80015c5a] : Timed out while waiting for systemd to remove kubepods-besteffort-pod510866f0_02ee_48ef_8219_fdea80015c5a.slice" Oct 04 10:53:00 crc kubenswrapper[5025]: I1004 10:53:00.782743 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5787d4b6b6-fc88x" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.457108 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-568f969787-rrbtk"] Oct 04 10:53:02 crc kubenswrapper[5025]: E1004 10:53:02.457647 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6622de76-40c3-434a-a87d-127e3d50fd4e" containerName="neutron-httpd" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.457659 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="6622de76-40c3-434a-a87d-127e3d50fd4e" containerName="neutron-httpd" Oct 04 10:53:02 crc kubenswrapper[5025]: E1004 10:53:02.457690 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6622de76-40c3-434a-a87d-127e3d50fd4e" containerName="neutron-api" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.457707 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="6622de76-40c3-434a-a87d-127e3d50fd4e" containerName="neutron-api" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.457865 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="6622de76-40c3-434a-a87d-127e3d50fd4e" containerName="neutron-api" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.457880 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="6622de76-40c3-434a-a87d-127e3d50fd4e" containerName="neutron-httpd" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.458754 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.463514 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-568f969787-rrbtk"] Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.464058 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.464572 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.464850 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.522763 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj7p5\" (UniqueName: \"kubernetes.io/projected/48087914-5acc-46e7-8ce2-382f6fe28d48-kube-api-access-rj7p5\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.522813 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-config-data\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.522870 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/48087914-5acc-46e7-8ce2-382f6fe28d48-etc-swift\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.522887 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-combined-ca-bundle\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.522904 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-internal-tls-certs\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.522932 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-public-tls-certs\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.522976 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48087914-5acc-46e7-8ce2-382f6fe28d48-run-httpd\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.523014 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48087914-5acc-46e7-8ce2-382f6fe28d48-log-httpd\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.624036 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj7p5\" (UniqueName: \"kubernetes.io/projected/48087914-5acc-46e7-8ce2-382f6fe28d48-kube-api-access-rj7p5\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.624092 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-config-data\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.624134 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-combined-ca-bundle\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.624151 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/48087914-5acc-46e7-8ce2-382f6fe28d48-etc-swift\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.624166 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-internal-tls-certs\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.624191 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-public-tls-certs\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.624225 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48087914-5acc-46e7-8ce2-382f6fe28d48-run-httpd\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.624251 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48087914-5acc-46e7-8ce2-382f6fe28d48-log-httpd\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.624874 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48087914-5acc-46e7-8ce2-382f6fe28d48-log-httpd\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.625118 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/48087914-5acc-46e7-8ce2-382f6fe28d48-run-httpd\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.630044 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-internal-tls-certs\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.630586 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-public-tls-certs\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.631772 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-combined-ca-bundle\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.632588 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48087914-5acc-46e7-8ce2-382f6fe28d48-config-data\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.634464 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/48087914-5acc-46e7-8ce2-382f6fe28d48-etc-swift\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.642347 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj7p5\" (UniqueName: \"kubernetes.io/projected/48087914-5acc-46e7-8ce2-382f6fe28d48-kube-api-access-rj7p5\") pod \"swift-proxy-568f969787-rrbtk\" (UID: \"48087914-5acc-46e7-8ce2-382f6fe28d48\") " pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:02 crc kubenswrapper[5025]: I1004 10:53:02.779438 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.359252 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-568f969787-rrbtk"] Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.411849 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.413349 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.417299 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-6b8c6" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.417654 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.417695 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.428818 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.438910 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.438994 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config-secret\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.439188 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6fc9\" (UniqueName: \"kubernetes.io/projected/cef6c557-e79c-487b-89da-4ece916caa08-kube-api-access-f6fc9\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.439209 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.540345 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.540435 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config-secret\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.540522 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6fc9\" (UniqueName: \"kubernetes.io/projected/cef6c557-e79c-487b-89da-4ece916caa08-kube-api-access-f6fc9\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.540545 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.541447 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.546789 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config-secret\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.549683 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-combined-ca-bundle\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.563175 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6fc9\" (UniqueName: \"kubernetes.io/projected/cef6c557-e79c-487b-89da-4ece916caa08-kube-api-access-f6fc9\") pod \"openstackclient\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.639984 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.640959 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.652670 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.680308 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.682077 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.686680 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.743916 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31297488-5835-4cd4-a9c5-7dff9253423a-openstack-config-secret\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.744018 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31297488-5835-4cd4-a9c5-7dff9253423a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.744099 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31297488-5835-4cd4-a9c5-7dff9253423a-openstack-config\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.744128 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rpdb\" (UniqueName: \"kubernetes.io/projected/31297488-5835-4cd4-a9c5-7dff9253423a-kube-api-access-8rpdb\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: E1004 10:53:03.749297 5025 log.go:32] "RunPodSandbox from runtime service failed" err=< Oct 04 10:53:03 crc kubenswrapper[5025]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_cef6c557-e79c-487b-89da-4ece916caa08_0(ce8b25ad096239b6165526246e209fbc27e5e13098bb695e33ded6c509042542): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ce8b25ad096239b6165526246e209fbc27e5e13098bb695e33ded6c509042542" Netns:"/var/run/netns/37cb4cde-1c85-4233-b5be-0883cf954115" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=ce8b25ad096239b6165526246e209fbc27e5e13098bb695e33ded6c509042542;K8S_POD_UID=cef6c557-e79c-487b-89da-4ece916caa08" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/cef6c557-e79c-487b-89da-4ece916caa08]: expected pod UID "cef6c557-e79c-487b-89da-4ece916caa08" but got "31297488-5835-4cd4-a9c5-7dff9253423a" from Kube API Oct 04 10:53:03 crc kubenswrapper[5025]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 04 10:53:03 crc kubenswrapper[5025]: > Oct 04 10:53:03 crc kubenswrapper[5025]: E1004 10:53:03.749372 5025 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Oct 04 10:53:03 crc kubenswrapper[5025]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_cef6c557-e79c-487b-89da-4ece916caa08_0(ce8b25ad096239b6165526246e209fbc27e5e13098bb695e33ded6c509042542): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ce8b25ad096239b6165526246e209fbc27e5e13098bb695e33ded6c509042542" Netns:"/var/run/netns/37cb4cde-1c85-4233-b5be-0883cf954115" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=ce8b25ad096239b6165526246e209fbc27e5e13098bb695e33ded6c509042542;K8S_POD_UID=cef6c557-e79c-487b-89da-4ece916caa08" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/cef6c557-e79c-487b-89da-4ece916caa08]: expected pod UID "cef6c557-e79c-487b-89da-4ece916caa08" but got "31297488-5835-4cd4-a9c5-7dff9253423a" from Kube API Oct 04 10:53:03 crc kubenswrapper[5025]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 04 10:53:03 crc kubenswrapper[5025]: > pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.846162 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31297488-5835-4cd4-a9c5-7dff9253423a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.846241 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31297488-5835-4cd4-a9c5-7dff9253423a-openstack-config\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.846263 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rpdb\" (UniqueName: \"kubernetes.io/projected/31297488-5835-4cd4-a9c5-7dff9253423a-kube-api-access-8rpdb\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.847221 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/31297488-5835-4cd4-a9c5-7dff9253423a-openstack-config\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.847423 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31297488-5835-4cd4-a9c5-7dff9253423a-openstack-config-secret\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.850883 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31297488-5835-4cd4-a9c5-7dff9253423a-combined-ca-bundle\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.851946 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/31297488-5835-4cd4-a9c5-7dff9253423a-openstack-config-secret\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:03 crc kubenswrapper[5025]: I1004 10:53:03.866159 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rpdb\" (UniqueName: \"kubernetes.io/projected/31297488-5835-4cd4-a9c5-7dff9253423a-kube-api-access-8rpdb\") pod \"openstackclient\" (UID: \"31297488-5835-4cd4-a9c5-7dff9253423a\") " pod="openstack/openstackclient" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.111841 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.218413 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.219064 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-568f969787-rrbtk" event={"ID":"48087914-5acc-46e7-8ce2-382f6fe28d48","Type":"ContainerStarted","Data":"7edc2832a69395591644f328a9f95bcf3b0ae2f00fac3b8272c190160a58521b"} Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.219092 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-568f969787-rrbtk" event={"ID":"48087914-5acc-46e7-8ce2-382f6fe28d48","Type":"ContainerStarted","Data":"49e5035e97c9186dba854f89cf91157f300fea2e6319bbc087117a8f3a4eaecf"} Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.219107 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-568f969787-rrbtk" event={"ID":"48087914-5acc-46e7-8ce2-382f6fe28d48","Type":"ContainerStarted","Data":"39384ec90712ec116ea2bb18ad424f0defb6c64d6b74af2e7ce3301c0b656986"} Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.219312 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.219343 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.221540 5025 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="cef6c557-e79c-487b-89da-4ece916caa08" podUID="31297488-5835-4cd4-a9c5-7dff9253423a" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.237027 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.254494 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-568f969787-rrbtk" podStartSLOduration=2.254472934 podStartE2EDuration="2.254472934s" podCreationTimestamp="2025-10-04 10:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:53:04.238979105 +0000 UTC m=+1112.663945985" watchObservedRunningTime="2025-10-04 10:53:04.254472934 +0000 UTC m=+1112.679439804" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.374848 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-combined-ca-bundle\") pod \"cef6c557-e79c-487b-89da-4ece916caa08\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.375011 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config\") pod \"cef6c557-e79c-487b-89da-4ece916caa08\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.375111 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config-secret\") pod \"cef6c557-e79c-487b-89da-4ece916caa08\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.375139 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6fc9\" (UniqueName: \"kubernetes.io/projected/cef6c557-e79c-487b-89da-4ece916caa08-kube-api-access-f6fc9\") pod \"cef6c557-e79c-487b-89da-4ece916caa08\" (UID: \"cef6c557-e79c-487b-89da-4ece916caa08\") " Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.375671 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "cef6c557-e79c-487b-89da-4ece916caa08" (UID: "cef6c557-e79c-487b-89da-4ece916caa08"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.376905 5025 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.380435 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cef6c557-e79c-487b-89da-4ece916caa08" (UID: "cef6c557-e79c-487b-89da-4ece916caa08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.380462 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cef6c557-e79c-487b-89da-4ece916caa08-kube-api-access-f6fc9" (OuterVolumeSpecName: "kube-api-access-f6fc9") pod "cef6c557-e79c-487b-89da-4ece916caa08" (UID: "cef6c557-e79c-487b-89da-4ece916caa08"). InnerVolumeSpecName "kube-api-access-f6fc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.380544 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "cef6c557-e79c-487b-89da-4ece916caa08" (UID: "cef6c557-e79c-487b-89da-4ece916caa08"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.422680 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cef6c557-e79c-487b-89da-4ece916caa08" path="/var/lib/kubelet/pods/cef6c557-e79c-487b-89da-4ece916caa08/volumes" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.479023 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.479073 5025 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/cef6c557-e79c-487b-89da-4ece916caa08-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.479086 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6fc9\" (UniqueName: \"kubernetes.io/projected/cef6c557-e79c-487b-89da-4ece916caa08-kube-api-access-f6fc9\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:04 crc kubenswrapper[5025]: I1004 10:53:04.548597 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 04 10:53:04 crc kubenswrapper[5025]: W1004 10:53:04.552013 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31297488_5835_4cd4_a9c5_7dff9253423a.slice/crio-d28934cfdb824e7f28184899ae2cdf661198914b36f5f8c84b458c4037798276 WatchSource:0}: Error finding container d28934cfdb824e7f28184899ae2cdf661198914b36f5f8c84b458c4037798276: Status 404 returned error can't find the container with id d28934cfdb824e7f28184899ae2cdf661198914b36f5f8c84b458c4037798276 Oct 04 10:53:05 crc kubenswrapper[5025]: I1004 10:53:05.015575 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 04 10:53:05 crc kubenswrapper[5025]: I1004 10:53:05.228344 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"31297488-5835-4cd4-a9c5-7dff9253423a","Type":"ContainerStarted","Data":"d28934cfdb824e7f28184899ae2cdf661198914b36f5f8c84b458c4037798276"} Oct 04 10:53:05 crc kubenswrapper[5025]: I1004 10:53:05.228467 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 04 10:53:05 crc kubenswrapper[5025]: I1004 10:53:05.235583 5025 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="cef6c557-e79c-487b-89da-4ece916caa08" podUID="31297488-5835-4cd4-a9c5-7dff9253423a" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.149081 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.279140 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-sg-core-conf-yaml\") pod \"03dee37f-5b63-4b21-b401-f05b07d3b1af\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.279199 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-config-data\") pod \"03dee37f-5b63-4b21-b401-f05b07d3b1af\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.279314 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-run-httpd\") pod \"03dee37f-5b63-4b21-b401-f05b07d3b1af\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.279412 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbbdh\" (UniqueName: \"kubernetes.io/projected/03dee37f-5b63-4b21-b401-f05b07d3b1af-kube-api-access-wbbdh\") pod \"03dee37f-5b63-4b21-b401-f05b07d3b1af\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.279439 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-combined-ca-bundle\") pod \"03dee37f-5b63-4b21-b401-f05b07d3b1af\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.279474 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-scripts\") pod \"03dee37f-5b63-4b21-b401-f05b07d3b1af\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.279576 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-log-httpd\") pod \"03dee37f-5b63-4b21-b401-f05b07d3b1af\" (UID: \"03dee37f-5b63-4b21-b401-f05b07d3b1af\") " Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.279978 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "03dee37f-5b63-4b21-b401-f05b07d3b1af" (UID: "03dee37f-5b63-4b21-b401-f05b07d3b1af"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.280158 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "03dee37f-5b63-4b21-b401-f05b07d3b1af" (UID: "03dee37f-5b63-4b21-b401-f05b07d3b1af"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.280587 5025 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.280609 5025 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03dee37f-5b63-4b21-b401-f05b07d3b1af-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.299112 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-scripts" (OuterVolumeSpecName: "scripts") pod "03dee37f-5b63-4b21-b401-f05b07d3b1af" (UID: "03dee37f-5b63-4b21-b401-f05b07d3b1af"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.303280 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03dee37f-5b63-4b21-b401-f05b07d3b1af-kube-api-access-wbbdh" (OuterVolumeSpecName: "kube-api-access-wbbdh") pod "03dee37f-5b63-4b21-b401-f05b07d3b1af" (UID: "03dee37f-5b63-4b21-b401-f05b07d3b1af"). InnerVolumeSpecName "kube-api-access-wbbdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.305248 5025 generic.go:334] "Generic (PLEG): container finished" podID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerID="dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062" exitCode=137 Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.305290 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03dee37f-5b63-4b21-b401-f05b07d3b1af","Type":"ContainerDied","Data":"dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062"} Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.305317 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"03dee37f-5b63-4b21-b401-f05b07d3b1af","Type":"ContainerDied","Data":"748fad10349346b8ecb24a1a99bfd2ff129149110b4144764266f0f137b83244"} Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.305336 5025 scope.go:117] "RemoveContainer" containerID="dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.305491 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.322788 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "03dee37f-5b63-4b21-b401-f05b07d3b1af" (UID: "03dee37f-5b63-4b21-b401-f05b07d3b1af"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.338922 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03dee37f-5b63-4b21-b401-f05b07d3b1af" (UID: "03dee37f-5b63-4b21-b401-f05b07d3b1af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.383166 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbbdh\" (UniqueName: \"kubernetes.io/projected/03dee37f-5b63-4b21-b401-f05b07d3b1af-kube-api-access-wbbdh\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.383637 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.383651 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.383666 5025 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.389061 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-config-data" (OuterVolumeSpecName: "config-data") pod "03dee37f-5b63-4b21-b401-f05b07d3b1af" (UID: "03dee37f-5b63-4b21-b401-f05b07d3b1af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.396473 5025 scope.go:117] "RemoveContainer" containerID="d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.417484 5025 scope.go:117] "RemoveContainer" containerID="872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.435997 5025 scope.go:117] "RemoveContainer" containerID="dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062" Oct 04 10:53:09 crc kubenswrapper[5025]: E1004 10:53:09.441111 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062\": container with ID starting with dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062 not found: ID does not exist" containerID="dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.441189 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062"} err="failed to get container status \"dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062\": rpc error: code = NotFound desc = could not find container \"dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062\": container with ID starting with dd6a088ce171cdc6510a430db999526a2cc4c2df04702a5064b523ff254c3062 not found: ID does not exist" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.441243 5025 scope.go:117] "RemoveContainer" containerID="d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230" Oct 04 10:53:09 crc kubenswrapper[5025]: E1004 10:53:09.441683 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230\": container with ID starting with d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230 not found: ID does not exist" containerID="d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.441719 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230"} err="failed to get container status \"d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230\": rpc error: code = NotFound desc = could not find container \"d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230\": container with ID starting with d4e1ec5ab90577411e2239e918bc7c42edca9ff1a8597eeff3ede2da29daa230 not found: ID does not exist" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.441740 5025 scope.go:117] "RemoveContainer" containerID="872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33" Oct 04 10:53:09 crc kubenswrapper[5025]: E1004 10:53:09.442104 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33\": container with ID starting with 872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33 not found: ID does not exist" containerID="872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.442125 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33"} err="failed to get container status \"872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33\": rpc error: code = NotFound desc = could not find container \"872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33\": container with ID starting with 872f73b9c531d386ec8cb8d1fe249f635a8f7f1b60da900404fc2b15526bbb33 not found: ID does not exist" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.486640 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03dee37f-5b63-4b21-b401-f05b07d3b1af-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.686502 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.696245 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.710106 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:09 crc kubenswrapper[5025]: E1004 10:53:09.710517 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="sg-core" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.710535 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="sg-core" Oct 04 10:53:09 crc kubenswrapper[5025]: E1004 10:53:09.710550 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="ceilometer-notification-agent" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.710559 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="ceilometer-notification-agent" Oct 04 10:53:09 crc kubenswrapper[5025]: E1004 10:53:09.710593 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="proxy-httpd" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.710599 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="proxy-httpd" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.710764 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="ceilometer-notification-agent" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.710784 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="sg-core" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.710804 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" containerName="proxy-httpd" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.712353 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.717092 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.718329 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.718328 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.792330 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-run-httpd\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.792414 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-log-httpd\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.792471 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.792862 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-config-data\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.793077 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.793307 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-scripts\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.793519 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkfzx\" (UniqueName: \"kubernetes.io/projected/bf0ddb7f-dad7-4c6f-81b3-300e24603527-kube-api-access-tkfzx\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.895560 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkfzx\" (UniqueName: \"kubernetes.io/projected/bf0ddb7f-dad7-4c6f-81b3-300e24603527-kube-api-access-tkfzx\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.895653 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-run-httpd\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.895679 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.895696 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-log-httpd\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.895749 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-config-data\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.895800 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.895863 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-scripts\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.896165 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-run-httpd\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.896502 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-log-httpd\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.900413 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-scripts\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.900545 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.901737 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.913426 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkfzx\" (UniqueName: \"kubernetes.io/projected/bf0ddb7f-dad7-4c6f-81b3-300e24603527-kube-api-access-tkfzx\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:09 crc kubenswrapper[5025]: I1004 10:53:09.917435 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-config-data\") pod \"ceilometer-0\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " pod="openstack/ceilometer-0" Oct 04 10:53:10 crc kubenswrapper[5025]: I1004 10:53:10.047879 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:10 crc kubenswrapper[5025]: I1004 10:53:10.422320 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03dee37f-5b63-4b21-b401-f05b07d3b1af" path="/var/lib/kubelet/pods/03dee37f-5b63-4b21-b401-f05b07d3b1af/volumes" Oct 04 10:53:10 crc kubenswrapper[5025]: I1004 10:53:10.782930 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5787d4b6b6-fc88x" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 10:53:12 crc kubenswrapper[5025]: I1004 10:53:12.792788 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:12 crc kubenswrapper[5025]: I1004 10:53:12.793185 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-568f969787-rrbtk" Oct 04 10:53:12 crc kubenswrapper[5025]: I1004 10:53:12.899597 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:16 crc kubenswrapper[5025]: W1004 10:53:16.106283 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf0ddb7f_dad7_4c6f_81b3_300e24603527.slice/crio-88285e63cd7ad347c0cb1bd4751cb251cb99cb907ad9d742327e17ca56b7159e WatchSource:0}: Error finding container 88285e63cd7ad347c0cb1bd4751cb251cb99cb907ad9d742327e17ca56b7159e: Status 404 returned error can't find the container with id 88285e63cd7ad347c0cb1bd4751cb251cb99cb907ad9d742327e17ca56b7159e Oct 04 10:53:16 crc kubenswrapper[5025]: I1004 10:53:16.108707 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:16 crc kubenswrapper[5025]: I1004 10:53:16.385505 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf0ddb7f-dad7-4c6f-81b3-300e24603527","Type":"ContainerStarted","Data":"88285e63cd7ad347c0cb1bd4751cb251cb99cb907ad9d742327e17ca56b7159e"} Oct 04 10:53:16 crc kubenswrapper[5025]: I1004 10:53:16.387794 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"31297488-5835-4cd4-a9c5-7dff9253423a","Type":"ContainerStarted","Data":"b1cb17c413d0859bdc82902704d924e000363f7ed1bc481fe551533b3f5895ee"} Oct 04 10:53:16 crc kubenswrapper[5025]: I1004 10:53:16.411320 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.35004085 podStartE2EDuration="13.411301494s" podCreationTimestamp="2025-10-04 10:53:03 +0000 UTC" firstStartedPulling="2025-10-04 10:53:04.554102094 +0000 UTC m=+1112.979068984" lastFinishedPulling="2025-10-04 10:53:15.615362748 +0000 UTC m=+1124.040329628" observedRunningTime="2025-10-04 10:53:16.40607323 +0000 UTC m=+1124.831040120" watchObservedRunningTime="2025-10-04 10:53:16.411301494 +0000 UTC m=+1124.836268374" Oct 04 10:53:17 crc kubenswrapper[5025]: I1004 10:53:17.397841 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf0ddb7f-dad7-4c6f-81b3-300e24603527","Type":"ContainerStarted","Data":"e53d8483c2884c055524d090e5b8772c915f800788d0021ffcf6009c79ef1850"} Oct 04 10:53:18 crc kubenswrapper[5025]: I1004 10:53:18.407186 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf0ddb7f-dad7-4c6f-81b3-300e24603527","Type":"ContainerStarted","Data":"2d803980063e485662a573a487fbcea735374b79f543c16e91abc8707f72e5da"} Oct 04 10:53:18 crc kubenswrapper[5025]: I1004 10:53:18.407499 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf0ddb7f-dad7-4c6f-81b3-300e24603527","Type":"ContainerStarted","Data":"a6e68b46ab503b58e973513a07e7dff5df9bd5d0f10d60bb2883f6099d63c724"} Oct 04 10:53:19 crc kubenswrapper[5025]: I1004 10:53:19.418279 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf0ddb7f-dad7-4c6f-81b3-300e24603527","Type":"ContainerStarted","Data":"9f72af4459f556a10bd73a6475f3326ad6b8caef952814c48a5e5342722a75e8"} Oct 04 10:53:19 crc kubenswrapper[5025]: I1004 10:53:19.418842 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 10:53:19 crc kubenswrapper[5025]: I1004 10:53:19.418675 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="proxy-httpd" containerID="cri-o://9f72af4459f556a10bd73a6475f3326ad6b8caef952814c48a5e5342722a75e8" gracePeriod=30 Oct 04 10:53:19 crc kubenswrapper[5025]: I1004 10:53:19.418419 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="ceilometer-central-agent" containerID="cri-o://e53d8483c2884c055524d090e5b8772c915f800788d0021ffcf6009c79ef1850" gracePeriod=30 Oct 04 10:53:19 crc kubenswrapper[5025]: I1004 10:53:19.418697 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="sg-core" containerID="cri-o://a6e68b46ab503b58e973513a07e7dff5df9bd5d0f10d60bb2883f6099d63c724" gracePeriod=30 Oct 04 10:53:19 crc kubenswrapper[5025]: I1004 10:53:19.418710 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="ceilometer-notification-agent" containerID="cri-o://2d803980063e485662a573a487fbcea735374b79f543c16e91abc8707f72e5da" gracePeriod=30 Oct 04 10:53:19 crc kubenswrapper[5025]: I1004 10:53:19.443242 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=7.5296250239999996 podStartE2EDuration="10.443225403s" podCreationTimestamp="2025-10-04 10:53:09 +0000 UTC" firstStartedPulling="2025-10-04 10:53:16.108598519 +0000 UTC m=+1124.533565399" lastFinishedPulling="2025-10-04 10:53:19.022198888 +0000 UTC m=+1127.447165778" observedRunningTime="2025-10-04 10:53:19.437794493 +0000 UTC m=+1127.862761383" watchObservedRunningTime="2025-10-04 10:53:19.443225403 +0000 UTC m=+1127.868192293" Oct 04 10:53:20 crc kubenswrapper[5025]: I1004 10:53:20.433814 5025 generic.go:334] "Generic (PLEG): container finished" podID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerID="9f72af4459f556a10bd73a6475f3326ad6b8caef952814c48a5e5342722a75e8" exitCode=0 Oct 04 10:53:20 crc kubenswrapper[5025]: I1004 10:53:20.433895 5025 generic.go:334] "Generic (PLEG): container finished" podID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerID="a6e68b46ab503b58e973513a07e7dff5df9bd5d0f10d60bb2883f6099d63c724" exitCode=2 Oct 04 10:53:20 crc kubenswrapper[5025]: I1004 10:53:20.433907 5025 generic.go:334] "Generic (PLEG): container finished" podID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerID="2d803980063e485662a573a487fbcea735374b79f543c16e91abc8707f72e5da" exitCode=0 Oct 04 10:53:20 crc kubenswrapper[5025]: I1004 10:53:20.433875 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf0ddb7f-dad7-4c6f-81b3-300e24603527","Type":"ContainerDied","Data":"9f72af4459f556a10bd73a6475f3326ad6b8caef952814c48a5e5342722a75e8"} Oct 04 10:53:20 crc kubenswrapper[5025]: I1004 10:53:20.436356 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf0ddb7f-dad7-4c6f-81b3-300e24603527","Type":"ContainerDied","Data":"a6e68b46ab503b58e973513a07e7dff5df9bd5d0f10d60bb2883f6099d63c724"} Oct 04 10:53:20 crc kubenswrapper[5025]: I1004 10:53:20.436391 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf0ddb7f-dad7-4c6f-81b3-300e24603527","Type":"ContainerDied","Data":"2d803980063e485662a573a487fbcea735374b79f543c16e91abc8707f72e5da"} Oct 04 10:53:20 crc kubenswrapper[5025]: I1004 10:53:20.782485 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5787d4b6b6-fc88x" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 04 10:53:20 crc kubenswrapper[5025]: I1004 10:53:20.782927 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:53:21 crc kubenswrapper[5025]: I1004 10:53:21.812056 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-t6kjx"] Oct 04 10:53:21 crc kubenswrapper[5025]: I1004 10:53:21.814720 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t6kjx" Oct 04 10:53:21 crc kubenswrapper[5025]: I1004 10:53:21.821279 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-t6kjx"] Oct 04 10:53:21 crc kubenswrapper[5025]: I1004 10:53:21.894469 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-9lvlk"] Oct 04 10:53:21 crc kubenswrapper[5025]: I1004 10:53:21.895539 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9lvlk" Oct 04 10:53:21 crc kubenswrapper[5025]: I1004 10:53:21.904126 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-9lvlk"] Oct 04 10:53:21 crc kubenswrapper[5025]: I1004 10:53:21.938650 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5mmw\" (UniqueName: \"kubernetes.io/projected/64a21479-f44b-49c1-bc9e-fafd77a4cf64-kube-api-access-x5mmw\") pod \"nova-api-db-create-t6kjx\" (UID: \"64a21479-f44b-49c1-bc9e-fafd77a4cf64\") " pod="openstack/nova-api-db-create-t6kjx" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.034007 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-bgvx5"] Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.035143 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bgvx5" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.040226 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5mmw\" (UniqueName: \"kubernetes.io/projected/64a21479-f44b-49c1-bc9e-fafd77a4cf64-kube-api-access-x5mmw\") pod \"nova-api-db-create-t6kjx\" (UID: \"64a21479-f44b-49c1-bc9e-fafd77a4cf64\") " pod="openstack/nova-api-db-create-t6kjx" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.040674 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t5db\" (UniqueName: \"kubernetes.io/projected/80e8d612-99b0-4396-a8dd-53b8de8956ba-kube-api-access-7t5db\") pod \"nova-cell0-db-create-9lvlk\" (UID: \"80e8d612-99b0-4396-a8dd-53b8de8956ba\") " pod="openstack/nova-cell0-db-create-9lvlk" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.046352 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bgvx5"] Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.092334 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5mmw\" (UniqueName: \"kubernetes.io/projected/64a21479-f44b-49c1-bc9e-fafd77a4cf64-kube-api-access-x5mmw\") pod \"nova-api-db-create-t6kjx\" (UID: \"64a21479-f44b-49c1-bc9e-fafd77a4cf64\") " pod="openstack/nova-api-db-create-t6kjx" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.141732 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t6kjx" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.142338 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27wvk\" (UniqueName: \"kubernetes.io/projected/ebdb7464-8cf2-439a-8789-e8b7617517fc-kube-api-access-27wvk\") pod \"nova-cell1-db-create-bgvx5\" (UID: \"ebdb7464-8cf2-439a-8789-e8b7617517fc\") " pod="openstack/nova-cell1-db-create-bgvx5" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.142595 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t5db\" (UniqueName: \"kubernetes.io/projected/80e8d612-99b0-4396-a8dd-53b8de8956ba-kube-api-access-7t5db\") pod \"nova-cell0-db-create-9lvlk\" (UID: \"80e8d612-99b0-4396-a8dd-53b8de8956ba\") " pod="openstack/nova-cell0-db-create-9lvlk" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.163162 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t5db\" (UniqueName: \"kubernetes.io/projected/80e8d612-99b0-4396-a8dd-53b8de8956ba-kube-api-access-7t5db\") pod \"nova-cell0-db-create-9lvlk\" (UID: \"80e8d612-99b0-4396-a8dd-53b8de8956ba\") " pod="openstack/nova-cell0-db-create-9lvlk" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.213925 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9lvlk" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.247384 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27wvk\" (UniqueName: \"kubernetes.io/projected/ebdb7464-8cf2-439a-8789-e8b7617517fc-kube-api-access-27wvk\") pod \"nova-cell1-db-create-bgvx5\" (UID: \"ebdb7464-8cf2-439a-8789-e8b7617517fc\") " pod="openstack/nova-cell1-db-create-bgvx5" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.272779 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27wvk\" (UniqueName: \"kubernetes.io/projected/ebdb7464-8cf2-439a-8789-e8b7617517fc-kube-api-access-27wvk\") pod \"nova-cell1-db-create-bgvx5\" (UID: \"ebdb7464-8cf2-439a-8789-e8b7617517fc\") " pod="openstack/nova-cell1-db-create-bgvx5" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.352847 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bgvx5" Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.691273 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-t6kjx"] Oct 04 10:53:22 crc kubenswrapper[5025]: W1004 10:53:22.691643 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64a21479_f44b_49c1_bc9e_fafd77a4cf64.slice/crio-db3a0ca60e4e0d0f71dd6bd87af59a490ac9e8490e7e3fd9ede8918dba85cf6c WatchSource:0}: Error finding container db3a0ca60e4e0d0f71dd6bd87af59a490ac9e8490e7e3fd9ede8918dba85cf6c: Status 404 returned error can't find the container with id db3a0ca60e4e0d0f71dd6bd87af59a490ac9e8490e7e3fd9ede8918dba85cf6c Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.753865 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-9lvlk"] Oct 04 10:53:22 crc kubenswrapper[5025]: I1004 10:53:22.856171 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bgvx5"] Oct 04 10:53:23 crc kubenswrapper[5025]: I1004 10:53:23.467483 5025 generic.go:334] "Generic (PLEG): container finished" podID="64a21479-f44b-49c1-bc9e-fafd77a4cf64" containerID="599e47f8c41209262f3f01ba560a31130c548fa284490bdd451e9a6e8cfd336c" exitCode=0 Oct 04 10:53:23 crc kubenswrapper[5025]: I1004 10:53:23.467673 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t6kjx" event={"ID":"64a21479-f44b-49c1-bc9e-fafd77a4cf64","Type":"ContainerDied","Data":"599e47f8c41209262f3f01ba560a31130c548fa284490bdd451e9a6e8cfd336c"} Oct 04 10:53:23 crc kubenswrapper[5025]: I1004 10:53:23.467790 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t6kjx" event={"ID":"64a21479-f44b-49c1-bc9e-fafd77a4cf64","Type":"ContainerStarted","Data":"db3a0ca60e4e0d0f71dd6bd87af59a490ac9e8490e7e3fd9ede8918dba85cf6c"} Oct 04 10:53:23 crc kubenswrapper[5025]: I1004 10:53:23.469407 5025 generic.go:334] "Generic (PLEG): container finished" podID="ebdb7464-8cf2-439a-8789-e8b7617517fc" containerID="32fefd7216c7fdfb5e4b657a7a7bd7a88d0a166619de4323b1694234028874cb" exitCode=0 Oct 04 10:53:23 crc kubenswrapper[5025]: I1004 10:53:23.469473 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bgvx5" event={"ID":"ebdb7464-8cf2-439a-8789-e8b7617517fc","Type":"ContainerDied","Data":"32fefd7216c7fdfb5e4b657a7a7bd7a88d0a166619de4323b1694234028874cb"} Oct 04 10:53:23 crc kubenswrapper[5025]: I1004 10:53:23.469706 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bgvx5" event={"ID":"ebdb7464-8cf2-439a-8789-e8b7617517fc","Type":"ContainerStarted","Data":"c7e68d4ae7313e937f19777a9eec1a94874b830f2fed462e61557ef7e52dcfe0"} Oct 04 10:53:23 crc kubenswrapper[5025]: I1004 10:53:23.471050 5025 generic.go:334] "Generic (PLEG): container finished" podID="80e8d612-99b0-4396-a8dd-53b8de8956ba" containerID="bc5c258a101b58a8e965b16b9856cad0d3faa7e793363769df66f4228fa6f4ea" exitCode=0 Oct 04 10:53:23 crc kubenswrapper[5025]: I1004 10:53:23.471088 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-9lvlk" event={"ID":"80e8d612-99b0-4396-a8dd-53b8de8956ba","Type":"ContainerDied","Data":"bc5c258a101b58a8e965b16b9856cad0d3faa7e793363769df66f4228fa6f4ea"} Oct 04 10:53:23 crc kubenswrapper[5025]: I1004 10:53:23.471111 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-9lvlk" event={"ID":"80e8d612-99b0-4396-a8dd-53b8de8956ba","Type":"ContainerStarted","Data":"751546ddcc9e949466461967e7e468f7e918d2bcb01f9cecf9e7c2297d5cdf12"} Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.502963 5025 generic.go:334] "Generic (PLEG): container finished" podID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerID="e53d8483c2884c055524d090e5b8772c915f800788d0021ffcf6009c79ef1850" exitCode=0 Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.503686 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf0ddb7f-dad7-4c6f-81b3-300e24603527","Type":"ContainerDied","Data":"e53d8483c2884c055524d090e5b8772c915f800788d0021ffcf6009c79ef1850"} Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.654924 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.796988 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-sg-core-conf-yaml\") pod \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.797093 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-log-httpd\") pod \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.797428 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bf0ddb7f-dad7-4c6f-81b3-300e24603527" (UID: "bf0ddb7f-dad7-4c6f-81b3-300e24603527"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.797493 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-run-httpd\") pod \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.797555 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-combined-ca-bundle\") pod \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.797592 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-config-data\") pod \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.797620 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-scripts\") pod \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.797642 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkfzx\" (UniqueName: \"kubernetes.io/projected/bf0ddb7f-dad7-4c6f-81b3-300e24603527-kube-api-access-tkfzx\") pod \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\" (UID: \"bf0ddb7f-dad7-4c6f-81b3-300e24603527\") " Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.798105 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bf0ddb7f-dad7-4c6f-81b3-300e24603527" (UID: "bf0ddb7f-dad7-4c6f-81b3-300e24603527"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.798283 5025 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.798297 5025 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bf0ddb7f-dad7-4c6f-81b3-300e24603527-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.803474 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf0ddb7f-dad7-4c6f-81b3-300e24603527-kube-api-access-tkfzx" (OuterVolumeSpecName: "kube-api-access-tkfzx") pod "bf0ddb7f-dad7-4c6f-81b3-300e24603527" (UID: "bf0ddb7f-dad7-4c6f-81b3-300e24603527"). InnerVolumeSpecName "kube-api-access-tkfzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.814988 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-scripts" (OuterVolumeSpecName: "scripts") pod "bf0ddb7f-dad7-4c6f-81b3-300e24603527" (UID: "bf0ddb7f-dad7-4c6f-81b3-300e24603527"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.845927 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bf0ddb7f-dad7-4c6f-81b3-300e24603527" (UID: "bf0ddb7f-dad7-4c6f-81b3-300e24603527"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.899523 5025 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.899561 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.899575 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkfzx\" (UniqueName: \"kubernetes.io/projected/bf0ddb7f-dad7-4c6f-81b3-300e24603527-kube-api-access-tkfzx\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.899617 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t6kjx" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.914335 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9lvlk" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.922650 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bgvx5" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.958413 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf0ddb7f-dad7-4c6f-81b3-300e24603527" (UID: "bf0ddb7f-dad7-4c6f-81b3-300e24603527"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:24 crc kubenswrapper[5025]: I1004 10:53:24.979163 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-config-data" (OuterVolumeSpecName: "config-data") pod "bf0ddb7f-dad7-4c6f-81b3-300e24603527" (UID: "bf0ddb7f-dad7-4c6f-81b3-300e24603527"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.000842 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5mmw\" (UniqueName: \"kubernetes.io/projected/64a21479-f44b-49c1-bc9e-fafd77a4cf64-kube-api-access-x5mmw\") pod \"64a21479-f44b-49c1-bc9e-fafd77a4cf64\" (UID: \"64a21479-f44b-49c1-bc9e-fafd77a4cf64\") " Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.001248 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.001264 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf0ddb7f-dad7-4c6f-81b3-300e24603527-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.004846 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64a21479-f44b-49c1-bc9e-fafd77a4cf64-kube-api-access-x5mmw" (OuterVolumeSpecName: "kube-api-access-x5mmw") pod "64a21479-f44b-49c1-bc9e-fafd77a4cf64" (UID: "64a21479-f44b-49c1-bc9e-fafd77a4cf64"). InnerVolumeSpecName "kube-api-access-x5mmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.101923 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7t5db\" (UniqueName: \"kubernetes.io/projected/80e8d612-99b0-4396-a8dd-53b8de8956ba-kube-api-access-7t5db\") pod \"80e8d612-99b0-4396-a8dd-53b8de8956ba\" (UID: \"80e8d612-99b0-4396-a8dd-53b8de8956ba\") " Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.102032 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27wvk\" (UniqueName: \"kubernetes.io/projected/ebdb7464-8cf2-439a-8789-e8b7617517fc-kube-api-access-27wvk\") pod \"ebdb7464-8cf2-439a-8789-e8b7617517fc\" (UID: \"ebdb7464-8cf2-439a-8789-e8b7617517fc\") " Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.102793 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5mmw\" (UniqueName: \"kubernetes.io/projected/64a21479-f44b-49c1-bc9e-fafd77a4cf64-kube-api-access-x5mmw\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.105195 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80e8d612-99b0-4396-a8dd-53b8de8956ba-kube-api-access-7t5db" (OuterVolumeSpecName: "kube-api-access-7t5db") pod "80e8d612-99b0-4396-a8dd-53b8de8956ba" (UID: "80e8d612-99b0-4396-a8dd-53b8de8956ba"). InnerVolumeSpecName "kube-api-access-7t5db". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.106598 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebdb7464-8cf2-439a-8789-e8b7617517fc-kube-api-access-27wvk" (OuterVolumeSpecName: "kube-api-access-27wvk") pod "ebdb7464-8cf2-439a-8789-e8b7617517fc" (UID: "ebdb7464-8cf2-439a-8789-e8b7617517fc"). InnerVolumeSpecName "kube-api-access-27wvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.204474 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7t5db\" (UniqueName: \"kubernetes.io/projected/80e8d612-99b0-4396-a8dd-53b8de8956ba-kube-api-access-7t5db\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.204516 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27wvk\" (UniqueName: \"kubernetes.io/projected/ebdb7464-8cf2-439a-8789-e8b7617517fc-kube-api-access-27wvk\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.521479 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-t6kjx" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.521474 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-t6kjx" event={"ID":"64a21479-f44b-49c1-bc9e-fafd77a4cf64","Type":"ContainerDied","Data":"db3a0ca60e4e0d0f71dd6bd87af59a490ac9e8490e7e3fd9ede8918dba85cf6c"} Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.522141 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db3a0ca60e4e0d0f71dd6bd87af59a490ac9e8490e7e3fd9ede8918dba85cf6c" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.524491 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bf0ddb7f-dad7-4c6f-81b3-300e24603527","Type":"ContainerDied","Data":"88285e63cd7ad347c0cb1bd4751cb251cb99cb907ad9d742327e17ca56b7159e"} Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.524534 5025 scope.go:117] "RemoveContainer" containerID="9f72af4459f556a10bd73a6475f3326ad6b8caef952814c48a5e5342722a75e8" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.524677 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.531788 5025 generic.go:334] "Generic (PLEG): container finished" podID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerID="16fdc70e3b1d7f1e5bbf6f4cd085a28452869621c0a13d1016341a4f70694f64" exitCode=137 Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.532163 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5787d4b6b6-fc88x" event={"ID":"1a4337f7-6dc7-4f98-8633-a61b61010693","Type":"ContainerDied","Data":"16fdc70e3b1d7f1e5bbf6f4cd085a28452869621c0a13d1016341a4f70694f64"} Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.532196 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5787d4b6b6-fc88x" event={"ID":"1a4337f7-6dc7-4f98-8633-a61b61010693","Type":"ContainerDied","Data":"db5d128e43585adecdb5dcfa84b1cd7d2595aafa9047d23a2c19faf08d09a728"} Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.532223 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db5d128e43585adecdb5dcfa84b1cd7d2595aafa9047d23a2c19faf08d09a728" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.534450 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-9lvlk" event={"ID":"80e8d612-99b0-4396-a8dd-53b8de8956ba","Type":"ContainerDied","Data":"751546ddcc9e949466461967e7e468f7e918d2bcb01f9cecf9e7c2297d5cdf12"} Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.534471 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="751546ddcc9e949466461967e7e468f7e918d2bcb01f9cecf9e7c2297d5cdf12" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.534522 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-9lvlk" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.537425 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bgvx5" event={"ID":"ebdb7464-8cf2-439a-8789-e8b7617517fc","Type":"ContainerDied","Data":"c7e68d4ae7313e937f19777a9eec1a94874b830f2fed462e61557ef7e52dcfe0"} Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.537459 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7e68d4ae7313e937f19777a9eec1a94874b830f2fed462e61557ef7e52dcfe0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.537488 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bgvx5" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.584401 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.586981 5025 scope.go:117] "RemoveContainer" containerID="a6e68b46ab503b58e973513a07e7dff5df9bd5d0f10d60bb2883f6099d63c724" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.621284 5025 scope.go:117] "RemoveContainer" containerID="2d803980063e485662a573a487fbcea735374b79f543c16e91abc8707f72e5da" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.645252 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.652747 5025 scope.go:117] "RemoveContainer" containerID="e53d8483c2884c055524d090e5b8772c915f800788d0021ffcf6009c79ef1850" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.665981 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.678594 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:25 crc kubenswrapper[5025]: E1004 10:53:25.679114 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="ceilometer-notification-agent" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679140 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="ceilometer-notification-agent" Oct 04 10:53:25 crc kubenswrapper[5025]: E1004 10:53:25.679173 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679183 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon" Oct 04 10:53:25 crc kubenswrapper[5025]: E1004 10:53:25.679198 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80e8d612-99b0-4396-a8dd-53b8de8956ba" containerName="mariadb-database-create" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679208 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="80e8d612-99b0-4396-a8dd-53b8de8956ba" containerName="mariadb-database-create" Oct 04 10:53:25 crc kubenswrapper[5025]: E1004 10:53:25.679221 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="proxy-httpd" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679229 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="proxy-httpd" Oct 04 10:53:25 crc kubenswrapper[5025]: E1004 10:53:25.679241 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon-log" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679249 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon-log" Oct 04 10:53:25 crc kubenswrapper[5025]: E1004 10:53:25.679267 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64a21479-f44b-49c1-bc9e-fafd77a4cf64" containerName="mariadb-database-create" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679276 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="64a21479-f44b-49c1-bc9e-fafd77a4cf64" containerName="mariadb-database-create" Oct 04 10:53:25 crc kubenswrapper[5025]: E1004 10:53:25.679287 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="ceilometer-central-agent" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679295 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="ceilometer-central-agent" Oct 04 10:53:25 crc kubenswrapper[5025]: E1004 10:53:25.679303 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebdb7464-8cf2-439a-8789-e8b7617517fc" containerName="mariadb-database-create" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679311 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebdb7464-8cf2-439a-8789-e8b7617517fc" containerName="mariadb-database-create" Oct 04 10:53:25 crc kubenswrapper[5025]: E1004 10:53:25.679322 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="sg-core" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679329 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="sg-core" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679560 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebdb7464-8cf2-439a-8789-e8b7617517fc" containerName="mariadb-database-create" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679592 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon-log" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679603 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="proxy-httpd" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679621 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" containerName="horizon" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679633 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="ceilometer-central-agent" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679646 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="ceilometer-notification-agent" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679657 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="64a21479-f44b-49c1-bc9e-fafd77a4cf64" containerName="mariadb-database-create" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679669 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" containerName="sg-core" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.679680 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="80e8d612-99b0-4396-a8dd-53b8de8956ba" containerName="mariadb-database-create" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.681760 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.683932 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.684232 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.691648 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.713411 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-secret-key\") pod \"1a4337f7-6dc7-4f98-8633-a61b61010693\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.713514 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8tx5\" (UniqueName: \"kubernetes.io/projected/1a4337f7-6dc7-4f98-8633-a61b61010693-kube-api-access-x8tx5\") pod \"1a4337f7-6dc7-4f98-8633-a61b61010693\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.713543 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-config-data\") pod \"1a4337f7-6dc7-4f98-8633-a61b61010693\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.713575 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-combined-ca-bundle\") pod \"1a4337f7-6dc7-4f98-8633-a61b61010693\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.713648 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-tls-certs\") pod \"1a4337f7-6dc7-4f98-8633-a61b61010693\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.713812 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4337f7-6dc7-4f98-8633-a61b61010693-logs\") pod \"1a4337f7-6dc7-4f98-8633-a61b61010693\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.713850 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-scripts\") pod \"1a4337f7-6dc7-4f98-8633-a61b61010693\" (UID: \"1a4337f7-6dc7-4f98-8633-a61b61010693\") " Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.716415 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a4337f7-6dc7-4f98-8633-a61b61010693-logs" (OuterVolumeSpecName: "logs") pod "1a4337f7-6dc7-4f98-8633-a61b61010693" (UID: "1a4337f7-6dc7-4f98-8633-a61b61010693"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.720196 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1a4337f7-6dc7-4f98-8633-a61b61010693" (UID: "1a4337f7-6dc7-4f98-8633-a61b61010693"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.726116 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a4337f7-6dc7-4f98-8633-a61b61010693-kube-api-access-x8tx5" (OuterVolumeSpecName: "kube-api-access-x8tx5") pod "1a4337f7-6dc7-4f98-8633-a61b61010693" (UID: "1a4337f7-6dc7-4f98-8633-a61b61010693"). InnerVolumeSpecName "kube-api-access-x8tx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.743580 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-scripts" (OuterVolumeSpecName: "scripts") pod "1a4337f7-6dc7-4f98-8633-a61b61010693" (UID: "1a4337f7-6dc7-4f98-8633-a61b61010693"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.745041 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-config-data" (OuterVolumeSpecName: "config-data") pod "1a4337f7-6dc7-4f98-8633-a61b61010693" (UID: "1a4337f7-6dc7-4f98-8633-a61b61010693"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.745596 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a4337f7-6dc7-4f98-8633-a61b61010693" (UID: "1a4337f7-6dc7-4f98-8633-a61b61010693"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.759884 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "1a4337f7-6dc7-4f98-8633-a61b61010693" (UID: "1a4337f7-6dc7-4f98-8633-a61b61010693"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.820274 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-log-httpd\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.820356 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-config-data\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.820489 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-run-httpd\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.820513 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.820575 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dstld\" (UniqueName: \"kubernetes.io/projected/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-kube-api-access-dstld\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.820655 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-scripts\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.820772 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.822005 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a4337f7-6dc7-4f98-8633-a61b61010693-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.822294 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.822370 5025 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.822430 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a4337f7-6dc7-4f98-8633-a61b61010693-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.822672 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8tx5\" (UniqueName: \"kubernetes.io/projected/1a4337f7-6dc7-4f98-8633-a61b61010693-kube-api-access-x8tx5\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.822736 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.822809 5025 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a4337f7-6dc7-4f98-8633-a61b61010693-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.925331 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-log-httpd\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.925398 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-config-data\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.925459 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-run-httpd\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.925483 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.925516 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dstld\" (UniqueName: \"kubernetes.io/projected/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-kube-api-access-dstld\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.925570 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-scripts\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.925626 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.927838 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-log-httpd\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.928130 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-run-httpd\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.933437 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.933467 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-scripts\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.933720 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.934953 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-config-data\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:25 crc kubenswrapper[5025]: I1004 10:53:25.950665 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dstld\" (UniqueName: \"kubernetes.io/projected/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-kube-api-access-dstld\") pod \"ceilometer-0\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " pod="openstack/ceilometer-0" Oct 04 10:53:26 crc kubenswrapper[5025]: I1004 10:53:26.020720 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:26 crc kubenswrapper[5025]: I1004 10:53:26.430474 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf0ddb7f-dad7-4c6f-81b3-300e24603527" path="/var/lib/kubelet/pods/bf0ddb7f-dad7-4c6f-81b3-300e24603527/volumes" Oct 04 10:53:26 crc kubenswrapper[5025]: I1004 10:53:26.460702 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:26 crc kubenswrapper[5025]: W1004 10:53:26.470394 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5928d69b_8d4a_45dc_9fe6_9b180fa577dc.slice/crio-3ed6561da9931bc4832d81dadae3d1c1678c0e833ae596f10dd496e157936844 WatchSource:0}: Error finding container 3ed6561da9931bc4832d81dadae3d1c1678c0e833ae596f10dd496e157936844: Status 404 returned error can't find the container with id 3ed6561da9931bc4832d81dadae3d1c1678c0e833ae596f10dd496e157936844 Oct 04 10:53:26 crc kubenswrapper[5025]: I1004 10:53:26.548285 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5928d69b-8d4a-45dc-9fe6-9b180fa577dc","Type":"ContainerStarted","Data":"3ed6561da9931bc4832d81dadae3d1c1678c0e833ae596f10dd496e157936844"} Oct 04 10:53:26 crc kubenswrapper[5025]: I1004 10:53:26.549400 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5787d4b6b6-fc88x" Oct 04 10:53:26 crc kubenswrapper[5025]: I1004 10:53:26.576556 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5787d4b6b6-fc88x"] Oct 04 10:53:26 crc kubenswrapper[5025]: I1004 10:53:26.583983 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5787d4b6b6-fc88x"] Oct 04 10:53:27 crc kubenswrapper[5025]: I1004 10:53:27.559439 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5928d69b-8d4a-45dc-9fe6-9b180fa577dc","Type":"ContainerStarted","Data":"2879bb26b9c0607a25ccec22658f5aaa795429d8fe44ffa9524e64d3b441a3ca"} Oct 04 10:53:28 crc kubenswrapper[5025]: I1004 10:53:28.245554 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:53:28 crc kubenswrapper[5025]: I1004 10:53:28.246252 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" containerName="glance-log" containerID="cri-o://1ae3a21b0cdb53e4bedf82ae70616497d5d1a0516a45bf2c84906ffe42903193" gracePeriod=30 Oct 04 10:53:28 crc kubenswrapper[5025]: I1004 10:53:28.246358 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" containerName="glance-httpd" containerID="cri-o://04134829f18094ec61789f563b5ffeba54b42c2452193c675283e042fa4241a2" gracePeriod=30 Oct 04 10:53:28 crc kubenswrapper[5025]: I1004 10:53:28.425528 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a4337f7-6dc7-4f98-8633-a61b61010693" path="/var/lib/kubelet/pods/1a4337f7-6dc7-4f98-8633-a61b61010693/volumes" Oct 04 10:53:28 crc kubenswrapper[5025]: I1004 10:53:28.571990 5025 generic.go:334] "Generic (PLEG): container finished" podID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" containerID="1ae3a21b0cdb53e4bedf82ae70616497d5d1a0516a45bf2c84906ffe42903193" exitCode=143 Oct 04 10:53:28 crc kubenswrapper[5025]: I1004 10:53:28.572068 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"773c73f6-a2c6-46c0-9616-a8b0782ad6bf","Type":"ContainerDied","Data":"1ae3a21b0cdb53e4bedf82ae70616497d5d1a0516a45bf2c84906ffe42903193"} Oct 04 10:53:28 crc kubenswrapper[5025]: I1004 10:53:28.576699 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5928d69b-8d4a-45dc-9fe6-9b180fa577dc","Type":"ContainerStarted","Data":"912308af5d66b48b288536d8aa3f1a8f1f13c2468145d497e1e95e5f3547c006"} Oct 04 10:53:29 crc kubenswrapper[5025]: I1004 10:53:29.082556 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:53:29 crc kubenswrapper[5025]: I1004 10:53:29.082816 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" containerName="glance-log" containerID="cri-o://9728653a1a8077c2d9e3b1c196c78e1f5847d3dc50e63f5e40634ec0f07be6d6" gracePeriod=30 Oct 04 10:53:29 crc kubenswrapper[5025]: I1004 10:53:29.082905 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" containerName="glance-httpd" containerID="cri-o://cfe3bfe62fe9e218a73048a8aef6b8aede522fd5277af0300f8f2796b80ac10b" gracePeriod=30 Oct 04 10:53:29 crc kubenswrapper[5025]: I1004 10:53:29.260925 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:29 crc kubenswrapper[5025]: I1004 10:53:29.591593 5025 generic.go:334] "Generic (PLEG): container finished" podID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" containerID="9728653a1a8077c2d9e3b1c196c78e1f5847d3dc50e63f5e40634ec0f07be6d6" exitCode=143 Oct 04 10:53:29 crc kubenswrapper[5025]: I1004 10:53:29.591699 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5","Type":"ContainerDied","Data":"9728653a1a8077c2d9e3b1c196c78e1f5847d3dc50e63f5e40634ec0f07be6d6"} Oct 04 10:53:29 crc kubenswrapper[5025]: I1004 10:53:29.610376 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5928d69b-8d4a-45dc-9fe6-9b180fa577dc","Type":"ContainerStarted","Data":"167ba4c8f31bdc3d228f74317ca3bbfa689cb372e585a25488aa64dac2ea15f5"} Oct 04 10:53:30 crc kubenswrapper[5025]: I1004 10:53:30.621127 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5928d69b-8d4a-45dc-9fe6-9b180fa577dc","Type":"ContainerStarted","Data":"10b1d62709fcf27cf9be7ca9eac1f18071521e47e747ff192c2b72577f5d50f5"} Oct 04 10:53:30 crc kubenswrapper[5025]: I1004 10:53:30.621286 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 10:53:30 crc kubenswrapper[5025]: I1004 10:53:30.621311 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="sg-core" containerID="cri-o://167ba4c8f31bdc3d228f74317ca3bbfa689cb372e585a25488aa64dac2ea15f5" gracePeriod=30 Oct 04 10:53:30 crc kubenswrapper[5025]: I1004 10:53:30.621639 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="ceilometer-notification-agent" containerID="cri-o://912308af5d66b48b288536d8aa3f1a8f1f13c2468145d497e1e95e5f3547c006" gracePeriod=30 Oct 04 10:53:30 crc kubenswrapper[5025]: I1004 10:53:30.621311 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="ceilometer-central-agent" containerID="cri-o://2879bb26b9c0607a25ccec22658f5aaa795429d8fe44ffa9524e64d3b441a3ca" gracePeriod=30 Oct 04 10:53:30 crc kubenswrapper[5025]: I1004 10:53:30.621803 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="proxy-httpd" containerID="cri-o://10b1d62709fcf27cf9be7ca9eac1f18071521e47e747ff192c2b72577f5d50f5" gracePeriod=30 Oct 04 10:53:30 crc kubenswrapper[5025]: I1004 10:53:30.648552 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.593014777 podStartE2EDuration="5.648536778s" podCreationTimestamp="2025-10-04 10:53:25 +0000 UTC" firstStartedPulling="2025-10-04 10:53:26.474046743 +0000 UTC m=+1134.899013623" lastFinishedPulling="2025-10-04 10:53:29.529568714 +0000 UTC m=+1137.954535624" observedRunningTime="2025-10-04 10:53:30.645309659 +0000 UTC m=+1139.070276539" watchObservedRunningTime="2025-10-04 10:53:30.648536778 +0000 UTC m=+1139.073503648" Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.633118 5025 generic.go:334] "Generic (PLEG): container finished" podID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" containerID="04134829f18094ec61789f563b5ffeba54b42c2452193c675283e042fa4241a2" exitCode=0 Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.633205 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"773c73f6-a2c6-46c0-9616-a8b0782ad6bf","Type":"ContainerDied","Data":"04134829f18094ec61789f563b5ffeba54b42c2452193c675283e042fa4241a2"} Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.635800 5025 generic.go:334] "Generic (PLEG): container finished" podID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerID="10b1d62709fcf27cf9be7ca9eac1f18071521e47e747ff192c2b72577f5d50f5" exitCode=0 Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.635819 5025 generic.go:334] "Generic (PLEG): container finished" podID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerID="167ba4c8f31bdc3d228f74317ca3bbfa689cb372e585a25488aa64dac2ea15f5" exitCode=2 Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.635826 5025 generic.go:334] "Generic (PLEG): container finished" podID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerID="912308af5d66b48b288536d8aa3f1a8f1f13c2468145d497e1e95e5f3547c006" exitCode=0 Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.635844 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5928d69b-8d4a-45dc-9fe6-9b180fa577dc","Type":"ContainerDied","Data":"10b1d62709fcf27cf9be7ca9eac1f18071521e47e747ff192c2b72577f5d50f5"} Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.635869 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5928d69b-8d4a-45dc-9fe6-9b180fa577dc","Type":"ContainerDied","Data":"167ba4c8f31bdc3d228f74317ca3bbfa689cb372e585a25488aa64dac2ea15f5"} Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.635878 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5928d69b-8d4a-45dc-9fe6-9b180fa577dc","Type":"ContainerDied","Data":"912308af5d66b48b288536d8aa3f1a8f1f13c2468145d497e1e95e5f3547c006"} Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.917657 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.945389 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-55f2-account-create-n2tgb"] Oct 04 10:53:31 crc kubenswrapper[5025]: E1004 10:53:31.945854 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" containerName="glance-log" Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.945872 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" containerName="glance-log" Oct 04 10:53:31 crc kubenswrapper[5025]: E1004 10:53:31.945886 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" containerName="glance-httpd" Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.945896 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" containerName="glance-httpd" Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.946164 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" containerName="glance-httpd" Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.946202 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" containerName="glance-log" Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.946914 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-55f2-account-create-n2tgb" Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.952947 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 04 10:53:31 crc kubenswrapper[5025]: I1004 10:53:31.956187 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-55f2-account-create-n2tgb"] Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.034904 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-config-data\") pod \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.034974 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-public-tls-certs\") pod \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.035062 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.035088 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-httpd-run\") pod \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.035113 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-scripts\") pod \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.035185 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-logs\") pod \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.035215 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49hr4\" (UniqueName: \"kubernetes.io/projected/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-kube-api-access-49hr4\") pod \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.035251 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-combined-ca-bundle\") pod \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\" (UID: \"773c73f6-a2c6-46c0-9616-a8b0782ad6bf\") " Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.035557 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcbqm\" (UniqueName: \"kubernetes.io/projected/603ede55-aaba-46e1-80fc-ba3ee57efa10-kube-api-access-zcbqm\") pod \"nova-api-55f2-account-create-n2tgb\" (UID: \"603ede55-aaba-46e1-80fc-ba3ee57efa10\") " pod="openstack/nova-api-55f2-account-create-n2tgb" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.036289 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-logs" (OuterVolumeSpecName: "logs") pod "773c73f6-a2c6-46c0-9616-a8b0782ad6bf" (UID: "773c73f6-a2c6-46c0-9616-a8b0782ad6bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.036467 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "773c73f6-a2c6-46c0-9616-a8b0782ad6bf" (UID: "773c73f6-a2c6-46c0-9616-a8b0782ad6bf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.041026 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-scripts" (OuterVolumeSpecName: "scripts") pod "773c73f6-a2c6-46c0-9616-a8b0782ad6bf" (UID: "773c73f6-a2c6-46c0-9616-a8b0782ad6bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.041540 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "773c73f6-a2c6-46c0-9616-a8b0782ad6bf" (UID: "773c73f6-a2c6-46c0-9616-a8b0782ad6bf"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.062725 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-kube-api-access-49hr4" (OuterVolumeSpecName: "kube-api-access-49hr4") pod "773c73f6-a2c6-46c0-9616-a8b0782ad6bf" (UID: "773c73f6-a2c6-46c0-9616-a8b0782ad6bf"). InnerVolumeSpecName "kube-api-access-49hr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.077340 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "773c73f6-a2c6-46c0-9616-a8b0782ad6bf" (UID: "773c73f6-a2c6-46c0-9616-a8b0782ad6bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.087290 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "773c73f6-a2c6-46c0-9616-a8b0782ad6bf" (UID: "773c73f6-a2c6-46c0-9616-a8b0782ad6bf"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.102395 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-config-data" (OuterVolumeSpecName: "config-data") pod "773c73f6-a2c6-46c0-9616-a8b0782ad6bf" (UID: "773c73f6-a2c6-46c0-9616-a8b0782ad6bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.141346 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcbqm\" (UniqueName: \"kubernetes.io/projected/603ede55-aaba-46e1-80fc-ba3ee57efa10-kube-api-access-zcbqm\") pod \"nova-api-55f2-account-create-n2tgb\" (UID: \"603ede55-aaba-46e1-80fc-ba3ee57efa10\") " pod="openstack/nova-api-55f2-account-create-n2tgb" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.141482 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.141499 5025 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.141539 5025 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.141551 5025 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.141564 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.145179 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.145263 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49hr4\" (UniqueName: \"kubernetes.io/projected/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-kube-api-access-49hr4\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.145313 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/773c73f6-a2c6-46c0-9616-a8b0782ad6bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.147925 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-4f45-account-create-xmrr8"] Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.149064 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4f45-account-create-xmrr8" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.150818 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.156811 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcbqm\" (UniqueName: \"kubernetes.io/projected/603ede55-aaba-46e1-80fc-ba3ee57efa10-kube-api-access-zcbqm\") pod \"nova-api-55f2-account-create-n2tgb\" (UID: \"603ede55-aaba-46e1-80fc-ba3ee57efa10\") " pod="openstack/nova-api-55f2-account-create-n2tgb" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.165067 5025 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.169094 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-4f45-account-create-xmrr8"] Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.246560 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qtnp\" (UniqueName: \"kubernetes.io/projected/7322f1a7-deea-4d50-a9c9-5cff160a4cbb-kube-api-access-5qtnp\") pod \"nova-cell0-4f45-account-create-xmrr8\" (UID: \"7322f1a7-deea-4d50-a9c9-5cff160a4cbb\") " pod="openstack/nova-cell0-4f45-account-create-xmrr8" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.247375 5025 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.247519 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-6551-account-create-bwsv6"] Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.248936 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6551-account-create-bwsv6" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.250601 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.254462 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-6551-account-create-bwsv6"] Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.275610 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-55f2-account-create-n2tgb" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.348969 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qtnp\" (UniqueName: \"kubernetes.io/projected/7322f1a7-deea-4d50-a9c9-5cff160a4cbb-kube-api-access-5qtnp\") pod \"nova-cell0-4f45-account-create-xmrr8\" (UID: \"7322f1a7-deea-4d50-a9c9-5cff160a4cbb\") " pod="openstack/nova-cell0-4f45-account-create-xmrr8" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.349095 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7p27\" (UniqueName: \"kubernetes.io/projected/0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24-kube-api-access-l7p27\") pod \"nova-cell1-6551-account-create-bwsv6\" (UID: \"0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24\") " pod="openstack/nova-cell1-6551-account-create-bwsv6" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.373204 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qtnp\" (UniqueName: \"kubernetes.io/projected/7322f1a7-deea-4d50-a9c9-5cff160a4cbb-kube-api-access-5qtnp\") pod \"nova-cell0-4f45-account-create-xmrr8\" (UID: \"7322f1a7-deea-4d50-a9c9-5cff160a4cbb\") " pod="openstack/nova-cell0-4f45-account-create-xmrr8" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.452745 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7p27\" (UniqueName: \"kubernetes.io/projected/0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24-kube-api-access-l7p27\") pod \"nova-cell1-6551-account-create-bwsv6\" (UID: \"0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24\") " pod="openstack/nova-cell1-6551-account-create-bwsv6" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.472599 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7p27\" (UniqueName: \"kubernetes.io/projected/0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24-kube-api-access-l7p27\") pod \"nova-cell1-6551-account-create-bwsv6\" (UID: \"0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24\") " pod="openstack/nova-cell1-6551-account-create-bwsv6" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.514457 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4f45-account-create-xmrr8" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.565992 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6551-account-create-bwsv6" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.653417 5025 generic.go:334] "Generic (PLEG): container finished" podID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" containerID="cfe3bfe62fe9e218a73048a8aef6b8aede522fd5277af0300f8f2796b80ac10b" exitCode=0 Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.653486 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5","Type":"ContainerDied","Data":"cfe3bfe62fe9e218a73048a8aef6b8aede522fd5277af0300f8f2796b80ac10b"} Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.658918 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"773c73f6-a2c6-46c0-9616-a8b0782ad6bf","Type":"ContainerDied","Data":"bd28d8339d4021c8df3b131b9e1c817d7b5e0192e8664efea9de876e249e89b8"} Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.659036 5025 scope.go:117] "RemoveContainer" containerID="04134829f18094ec61789f563b5ffeba54b42c2452193c675283e042fa4241a2" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.659047 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.702437 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.725345 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.733702 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.736390 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.737485 5025 scope.go:117] "RemoveContainer" containerID="1ae3a21b0cdb53e4bedf82ae70616497d5d1a0516a45bf2c84906ffe42903193" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.742398 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.742623 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.742745 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.802250 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-55f2-account-create-n2tgb"] Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.830615 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.875793 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35af90f2-e78a-4e59-a694-72044d020d8f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.875858 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-scripts\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.875922 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.876049 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35af90f2-e78a-4e59-a694-72044d020d8f-logs\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.876090 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8w8r\" (UniqueName: \"kubernetes.io/projected/35af90f2-e78a-4e59-a694-72044d020d8f-kube-api-access-z8w8r\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.876128 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.876178 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-config-data\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.876234 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.940124 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.981186 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.981266 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-config-data\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.981319 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.981342 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35af90f2-e78a-4e59-a694-72044d020d8f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.981839 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/35af90f2-e78a-4e59-a694-72044d020d8f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.982214 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-scripts\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.982296 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.982400 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35af90f2-e78a-4e59-a694-72044d020d8f-logs\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.982449 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8w8r\" (UniqueName: \"kubernetes.io/projected/35af90f2-e78a-4e59-a694-72044d020d8f-kube-api-access-z8w8r\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.982870 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.982893 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35af90f2-e78a-4e59-a694-72044d020d8f-logs\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:32 crc kubenswrapper[5025]: I1004 10:53:32.992260 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-config-data\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.000695 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.001207 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-scripts\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.002932 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/35af90f2-e78a-4e59-a694-72044d020d8f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.005707 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8w8r\" (UniqueName: \"kubernetes.io/projected/35af90f2-e78a-4e59-a694-72044d020d8f-kube-api-access-z8w8r\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.052236 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-4f45-account-create-xmrr8"] Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.076550 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.078221 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"35af90f2-e78a-4e59-a694-72044d020d8f\") " pod="openstack/glance-default-external-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.083726 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.083793 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-config-data\") pod \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.083882 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-internal-tls-certs\") pod \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.083933 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-combined-ca-bundle\") pod \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.083992 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-httpd-run\") pod \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.084091 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-logs\") pod \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.084204 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49n6s\" (UniqueName: \"kubernetes.io/projected/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-kube-api-access-49n6s\") pod \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.084235 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-scripts\") pod \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\" (UID: \"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5\") " Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.088163 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" (UID: "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.089853 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-logs" (OuterVolumeSpecName: "logs") pod "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" (UID: "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.093145 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-scripts" (OuterVolumeSpecName: "scripts") pod "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" (UID: "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.095450 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-kube-api-access-49n6s" (OuterVolumeSpecName: "kube-api-access-49n6s") pod "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" (UID: "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5"). InnerVolumeSpecName "kube-api-access-49n6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.095784 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" (UID: "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.126238 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" (UID: "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.157378 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-config-data" (OuterVolumeSpecName: "config-data") pod "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" (UID: "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.170406 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-6551-account-create-bwsv6"] Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.186302 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49n6s\" (UniqueName: \"kubernetes.io/projected/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-kube-api-access-49n6s\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.186342 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.186372 5025 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.186386 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.186401 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.186412 5025 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.186423 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.187514 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.199580 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" (UID: "0ef75483-c407-4ba6-8cfe-9b7a663ce7f5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.211111 5025 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.289237 5025 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.289287 5025 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.378836 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.670279 5025 generic.go:334] "Generic (PLEG): container finished" podID="603ede55-aaba-46e1-80fc-ba3ee57efa10" containerID="0a25d1cb9587b5420387869af2bc1fa857e9c0c55125ac33f569b2ce792ced10" exitCode=0 Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.670435 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-55f2-account-create-n2tgb" event={"ID":"603ede55-aaba-46e1-80fc-ba3ee57efa10","Type":"ContainerDied","Data":"0a25d1cb9587b5420387869af2bc1fa857e9c0c55125ac33f569b2ce792ced10"} Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.670599 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-55f2-account-create-n2tgb" event={"ID":"603ede55-aaba-46e1-80fc-ba3ee57efa10","Type":"ContainerStarted","Data":"f8d86cbf1a6cf0990da140e3f06b26500dd125c11b833c2b0e6845ee8407436c"} Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.677486 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.679168 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"0ef75483-c407-4ba6-8cfe-9b7a663ce7f5","Type":"ContainerDied","Data":"1ccf6b52cc6995544a2732356a6ab15d6ec29829771f896f0c034fe9d6926570"} Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.679225 5025 scope.go:117] "RemoveContainer" containerID="cfe3bfe62fe9e218a73048a8aef6b8aede522fd5277af0300f8f2796b80ac10b" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.681417 5025 generic.go:334] "Generic (PLEG): container finished" podID="7322f1a7-deea-4d50-a9c9-5cff160a4cbb" containerID="119283c99064e440b346a12cc67ff3aa47f7ec86d892751d424358afe9eeaf5c" exitCode=0 Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.681463 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4f45-account-create-xmrr8" event={"ID":"7322f1a7-deea-4d50-a9c9-5cff160a4cbb","Type":"ContainerDied","Data":"119283c99064e440b346a12cc67ff3aa47f7ec86d892751d424358afe9eeaf5c"} Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.681479 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4f45-account-create-xmrr8" event={"ID":"7322f1a7-deea-4d50-a9c9-5cff160a4cbb","Type":"ContainerStarted","Data":"56e2d6aa2651da1271f093e69495322ba0bc61fdf40af861eb719960f0aafef2"} Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.695179 5025 generic.go:334] "Generic (PLEG): container finished" podID="0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24" containerID="015d8ba1eef6ebe0ebc1c43f6e6e47bf6cba242f756777f201743f97872fc491" exitCode=0 Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.695221 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6551-account-create-bwsv6" event={"ID":"0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24","Type":"ContainerDied","Data":"015d8ba1eef6ebe0ebc1c43f6e6e47bf6cba242f756777f201743f97872fc491"} Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.695244 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6551-account-create-bwsv6" event={"ID":"0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24","Type":"ContainerStarted","Data":"53e714ee3b6ab9bde0520dce7809cca86d3ca7b0b77e71f7534ea5d516b73018"} Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.710504 5025 scope.go:117] "RemoveContainer" containerID="9728653a1a8077c2d9e3b1c196c78e1f5847d3dc50e63f5e40634ec0f07be6d6" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.748867 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.758781 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.772339 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:53:33 crc kubenswrapper[5025]: E1004 10:53:33.772812 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" containerName="glance-httpd" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.772837 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" containerName="glance-httpd" Oct 04 10:53:33 crc kubenswrapper[5025]: E1004 10:53:33.772874 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" containerName="glance-log" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.772882 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" containerName="glance-log" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.773448 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" containerName="glance-log" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.773487 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" containerName="glance-httpd" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.775597 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.777562 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.778633 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.796134 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.900139 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.900227 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffh9t\" (UniqueName: \"kubernetes.io/projected/5393a1c0-fcc5-449d-a4e5-42723c13d83e-kube-api-access-ffh9t\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.900300 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.900338 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5393a1c0-fcc5-449d-a4e5-42723c13d83e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.900361 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.900402 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.900452 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5393a1c0-fcc5-449d-a4e5-42723c13d83e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.900474 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:33 crc kubenswrapper[5025]: I1004 10:53:33.918304 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.002524 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.002617 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffh9t\" (UniqueName: \"kubernetes.io/projected/5393a1c0-fcc5-449d-a4e5-42723c13d83e-kube-api-access-ffh9t\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.002688 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.002727 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.002746 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5393a1c0-fcc5-449d-a4e5-42723c13d83e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.002785 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.002836 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5393a1c0-fcc5-449d-a4e5-42723c13d83e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.002856 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.003485 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5393a1c0-fcc5-449d-a4e5-42723c13d83e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.003562 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5393a1c0-fcc5-449d-a4e5-42723c13d83e-logs\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.003850 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.009807 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.009942 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.010138 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.011821 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5393a1c0-fcc5-449d-a4e5-42723c13d83e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.024100 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffh9t\" (UniqueName: \"kubernetes.io/projected/5393a1c0-fcc5-449d-a4e5-42723c13d83e-kube-api-access-ffh9t\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.040356 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"5393a1c0-fcc5-449d-a4e5-42723c13d83e\") " pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.097846 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.422240 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ef75483-c407-4ba6-8cfe-9b7a663ce7f5" path="/var/lib/kubelet/pods/0ef75483-c407-4ba6-8cfe-9b7a663ce7f5/volumes" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.423475 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="773c73f6-a2c6-46c0-9616-a8b0782ad6bf" path="/var/lib/kubelet/pods/773c73f6-a2c6-46c0-9616-a8b0782ad6bf/volumes" Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.609098 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 04 10:53:34 crc kubenswrapper[5025]: W1004 10:53:34.614280 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5393a1c0_fcc5_449d_a4e5_42723c13d83e.slice/crio-e1bd5bcb72333785537ed90c669c8c7c3e4f409fb9700d27a7638309e77de3cf WatchSource:0}: Error finding container e1bd5bcb72333785537ed90c669c8c7c3e4f409fb9700d27a7638309e77de3cf: Status 404 returned error can't find the container with id e1bd5bcb72333785537ed90c669c8c7c3e4f409fb9700d27a7638309e77de3cf Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.724771 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"35af90f2-e78a-4e59-a694-72044d020d8f","Type":"ContainerStarted","Data":"3a44be594a595eb695d47e4a104bf22938ff33c2fa10539e78475a8f0568d492"} Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.724818 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"35af90f2-e78a-4e59-a694-72044d020d8f","Type":"ContainerStarted","Data":"f68c484dc536545db83483110bddd04fa4758ec66234808fe0a8d9f64c098cb0"} Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.730198 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5393a1c0-fcc5-449d-a4e5-42723c13d83e","Type":"ContainerStarted","Data":"e1bd5bcb72333785537ed90c669c8c7c3e4f409fb9700d27a7638309e77de3cf"} Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.733411 5025 generic.go:334] "Generic (PLEG): container finished" podID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerID="2879bb26b9c0607a25ccec22658f5aaa795429d8fe44ffa9524e64d3b441a3ca" exitCode=0 Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.733558 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5928d69b-8d4a-45dc-9fe6-9b180fa577dc","Type":"ContainerDied","Data":"2879bb26b9c0607a25ccec22658f5aaa795429d8fe44ffa9524e64d3b441a3ca"} Oct 04 10:53:34 crc kubenswrapper[5025]: I1004 10:53:34.935578 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.037630 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-run-httpd\") pod \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.037718 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-log-httpd\") pod \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.037752 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dstld\" (UniqueName: \"kubernetes.io/projected/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-kube-api-access-dstld\") pod \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.038563 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5928d69b-8d4a-45dc-9fe6-9b180fa577dc" (UID: "5928d69b-8d4a-45dc-9fe6-9b180fa577dc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.038652 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-combined-ca-bundle\") pod \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.039075 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5928d69b-8d4a-45dc-9fe6-9b180fa577dc" (UID: "5928d69b-8d4a-45dc-9fe6-9b180fa577dc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.039215 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-sg-core-conf-yaml\") pod \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.039296 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-scripts\") pod \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.039319 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-config-data\") pod \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\" (UID: \"5928d69b-8d4a-45dc-9fe6-9b180fa577dc\") " Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.039792 5025 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.039806 5025 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.075314 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-kube-api-access-dstld" (OuterVolumeSpecName: "kube-api-access-dstld") pod "5928d69b-8d4a-45dc-9fe6-9b180fa577dc" (UID: "5928d69b-8d4a-45dc-9fe6-9b180fa577dc"). InnerVolumeSpecName "kube-api-access-dstld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.075820 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-scripts" (OuterVolumeSpecName: "scripts") pod "5928d69b-8d4a-45dc-9fe6-9b180fa577dc" (UID: "5928d69b-8d4a-45dc-9fe6-9b180fa577dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.098137 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5928d69b-8d4a-45dc-9fe6-9b180fa577dc" (UID: "5928d69b-8d4a-45dc-9fe6-9b180fa577dc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.139965 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5928d69b-8d4a-45dc-9fe6-9b180fa577dc" (UID: "5928d69b-8d4a-45dc-9fe6-9b180fa577dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.140925 5025 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.140947 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.140955 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dstld\" (UniqueName: \"kubernetes.io/projected/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-kube-api-access-dstld\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.140966 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.187438 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-config-data" (OuterVolumeSpecName: "config-data") pod "5928d69b-8d4a-45dc-9fe6-9b180fa577dc" (UID: "5928d69b-8d4a-45dc-9fe6-9b180fa577dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.204551 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-55f2-account-create-n2tgb" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.242262 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5928d69b-8d4a-45dc-9fe6-9b180fa577dc-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.243006 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4f45-account-create-xmrr8" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.272859 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6551-account-create-bwsv6" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.343391 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7p27\" (UniqueName: \"kubernetes.io/projected/0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24-kube-api-access-l7p27\") pod \"0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24\" (UID: \"0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24\") " Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.343495 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qtnp\" (UniqueName: \"kubernetes.io/projected/7322f1a7-deea-4d50-a9c9-5cff160a4cbb-kube-api-access-5qtnp\") pod \"7322f1a7-deea-4d50-a9c9-5cff160a4cbb\" (UID: \"7322f1a7-deea-4d50-a9c9-5cff160a4cbb\") " Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.343579 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcbqm\" (UniqueName: \"kubernetes.io/projected/603ede55-aaba-46e1-80fc-ba3ee57efa10-kube-api-access-zcbqm\") pod \"603ede55-aaba-46e1-80fc-ba3ee57efa10\" (UID: \"603ede55-aaba-46e1-80fc-ba3ee57efa10\") " Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.346680 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24-kube-api-access-l7p27" (OuterVolumeSpecName: "kube-api-access-l7p27") pod "0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24" (UID: "0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24"). InnerVolumeSpecName "kube-api-access-l7p27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.347143 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7322f1a7-deea-4d50-a9c9-5cff160a4cbb-kube-api-access-5qtnp" (OuterVolumeSpecName: "kube-api-access-5qtnp") pod "7322f1a7-deea-4d50-a9c9-5cff160a4cbb" (UID: "7322f1a7-deea-4d50-a9c9-5cff160a4cbb"). InnerVolumeSpecName "kube-api-access-5qtnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.347330 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/603ede55-aaba-46e1-80fc-ba3ee57efa10-kube-api-access-zcbqm" (OuterVolumeSpecName: "kube-api-access-zcbqm") pod "603ede55-aaba-46e1-80fc-ba3ee57efa10" (UID: "603ede55-aaba-46e1-80fc-ba3ee57efa10"). InnerVolumeSpecName "kube-api-access-zcbqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.446208 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7p27\" (UniqueName: \"kubernetes.io/projected/0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24-kube-api-access-l7p27\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.446249 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qtnp\" (UniqueName: \"kubernetes.io/projected/7322f1a7-deea-4d50-a9c9-5cff160a4cbb-kube-api-access-5qtnp\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.446258 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcbqm\" (UniqueName: \"kubernetes.io/projected/603ede55-aaba-46e1-80fc-ba3ee57efa10-kube-api-access-zcbqm\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.747834 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5928d69b-8d4a-45dc-9fe6-9b180fa577dc","Type":"ContainerDied","Data":"3ed6561da9931bc4832d81dadae3d1c1678c0e833ae596f10dd496e157936844"} Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.748231 5025 scope.go:117] "RemoveContainer" containerID="10b1d62709fcf27cf9be7ca9eac1f18071521e47e747ff192c2b72577f5d50f5" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.747925 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.753198 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"35af90f2-e78a-4e59-a694-72044d020d8f","Type":"ContainerStarted","Data":"e12ea648d559d07d713f75526090072167b28ac0be10c2db5a522465b41f708e"} Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.756237 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-4f45-account-create-xmrr8" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.756240 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-4f45-account-create-xmrr8" event={"ID":"7322f1a7-deea-4d50-a9c9-5cff160a4cbb","Type":"ContainerDied","Data":"56e2d6aa2651da1271f093e69495322ba0bc61fdf40af861eb719960f0aafef2"} Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.756342 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56e2d6aa2651da1271f093e69495322ba0bc61fdf40af861eb719960f0aafef2" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.762960 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5393a1c0-fcc5-449d-a4e5-42723c13d83e","Type":"ContainerStarted","Data":"6e4abfa9a45487ec6d4cac0034ff45e1c526556fe6aea73b682c6d9dbb01faa4"} Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.764483 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-6551-account-create-bwsv6" event={"ID":"0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24","Type":"ContainerDied","Data":"53e714ee3b6ab9bde0520dce7809cca86d3ca7b0b77e71f7534ea5d516b73018"} Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.764513 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53e714ee3b6ab9bde0520dce7809cca86d3ca7b0b77e71f7534ea5d516b73018" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.764564 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-6551-account-create-bwsv6" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.781250 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.7812311530000002 podStartE2EDuration="3.781231153s" podCreationTimestamp="2025-10-04 10:53:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:53:35.773472219 +0000 UTC m=+1144.198439119" watchObservedRunningTime="2025-10-04 10:53:35.781231153 +0000 UTC m=+1144.206198033" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.792943 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-55f2-account-create-n2tgb" event={"ID":"603ede55-aaba-46e1-80fc-ba3ee57efa10","Type":"ContainerDied","Data":"f8d86cbf1a6cf0990da140e3f06b26500dd125c11b833c2b0e6845ee8407436c"} Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.793002 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8d86cbf1a6cf0990da140e3f06b26500dd125c11b833c2b0e6845ee8407436c" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.793090 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-55f2-account-create-n2tgb" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.803834 5025 scope.go:117] "RemoveContainer" containerID="167ba4c8f31bdc3d228f74317ca3bbfa689cb372e585a25488aa64dac2ea15f5" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.828116 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.841296 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853080 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:35 crc kubenswrapper[5025]: E1004 10:53:35.853457 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="ceilometer-central-agent" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853472 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="ceilometer-central-agent" Oct 04 10:53:35 crc kubenswrapper[5025]: E1004 10:53:35.853492 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24" containerName="mariadb-account-create" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853498 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24" containerName="mariadb-account-create" Oct 04 10:53:35 crc kubenswrapper[5025]: E1004 10:53:35.853507 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="sg-core" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853513 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="sg-core" Oct 04 10:53:35 crc kubenswrapper[5025]: E1004 10:53:35.853520 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="603ede55-aaba-46e1-80fc-ba3ee57efa10" containerName="mariadb-account-create" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853525 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="603ede55-aaba-46e1-80fc-ba3ee57efa10" containerName="mariadb-account-create" Oct 04 10:53:35 crc kubenswrapper[5025]: E1004 10:53:35.853539 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7322f1a7-deea-4d50-a9c9-5cff160a4cbb" containerName="mariadb-account-create" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853545 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="7322f1a7-deea-4d50-a9c9-5cff160a4cbb" containerName="mariadb-account-create" Oct 04 10:53:35 crc kubenswrapper[5025]: E1004 10:53:35.853557 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="proxy-httpd" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853563 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="proxy-httpd" Oct 04 10:53:35 crc kubenswrapper[5025]: E1004 10:53:35.853574 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="ceilometer-notification-agent" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853581 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="ceilometer-notification-agent" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853744 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="sg-core" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853759 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="7322f1a7-deea-4d50-a9c9-5cff160a4cbb" containerName="mariadb-account-create" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853770 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24" containerName="mariadb-account-create" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853783 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="ceilometer-central-agent" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853794 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="603ede55-aaba-46e1-80fc-ba3ee57efa10" containerName="mariadb-account-create" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853806 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="proxy-httpd" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.853816 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" containerName="ceilometer-notification-agent" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.855341 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.858415 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.858422 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.875199 5025 scope.go:117] "RemoveContainer" containerID="912308af5d66b48b288536d8aa3f1a8f1f13c2468145d497e1e95e5f3547c006" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.883783 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.936179 5025 scope.go:117] "RemoveContainer" containerID="2879bb26b9c0607a25ccec22658f5aaa795429d8fe44ffa9524e64d3b441a3ca" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.956775 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.956883 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-scripts\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.956999 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-config-data\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.959161 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.959329 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-log-httpd\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.959363 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg2c2\" (UniqueName: \"kubernetes.io/projected/2aee4089-d0b3-4caf-89a6-70902456cf61-kube-api-access-pg2c2\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:35 crc kubenswrapper[5025]: I1004 10:53:35.959401 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-run-httpd\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.061019 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-config-data\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.061162 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.061194 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-log-httpd\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.061247 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg2c2\" (UniqueName: \"kubernetes.io/projected/2aee4089-d0b3-4caf-89a6-70902456cf61-kube-api-access-pg2c2\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.061280 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-run-httpd\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.061338 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.061368 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-scripts\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.061666 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-log-httpd\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.061948 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-run-httpd\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.068762 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-scripts\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.069350 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.069863 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-config-data\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.080660 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg2c2\" (UniqueName: \"kubernetes.io/projected/2aee4089-d0b3-4caf-89a6-70902456cf61-kube-api-access-pg2c2\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.081569 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.225429 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.431725 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5928d69b-8d4a-45dc-9fe6-9b180fa577dc" path="/var/lib/kubelet/pods/5928d69b-8d4a-45dc-9fe6-9b180fa577dc/volumes" Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.712132 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.807856 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2aee4089-d0b3-4caf-89a6-70902456cf61","Type":"ContainerStarted","Data":"42ea50d7f3f4a4ca840aa0b40f62b04fad46a4e766c25c2b808169a9916fd480"} Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.811141 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5393a1c0-fcc5-449d-a4e5-42723c13d83e","Type":"ContainerStarted","Data":"12c85156e6b89ec3df0bf279a81f7ba1e9b8d77007e14f6dd198ad4e31327979"} Oct 04 10:53:36 crc kubenswrapper[5025]: I1004 10:53:36.840159 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.840141356 podStartE2EDuration="3.840141356s" podCreationTimestamp="2025-10-04 10:53:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:53:36.837922455 +0000 UTC m=+1145.262889335" watchObservedRunningTime="2025-10-04 10:53:36.840141356 +0000 UTC m=+1145.265108236" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.570767 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fdcd7"] Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.572797 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.579381 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vxxmc" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.579499 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.579555 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.586562 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fdcd7"] Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.693694 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-config-data\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.693863 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-scripts\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.693911 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd2ds\" (UniqueName: \"kubernetes.io/projected/70c3c42b-9625-43a5-9d21-a145d8380fc0-kube-api-access-wd2ds\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.693941 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.799490 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-config-data\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.799625 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-scripts\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.799651 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd2ds\" (UniqueName: \"kubernetes.io/projected/70c3c42b-9625-43a5-9d21-a145d8380fc0-kube-api-access-wd2ds\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.799680 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.805497 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-scripts\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.805718 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-config-data\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.815349 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd2ds\" (UniqueName: \"kubernetes.io/projected/70c3c42b-9625-43a5-9d21-a145d8380fc0-kube-api-access-wd2ds\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.816613 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fdcd7\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:37 crc kubenswrapper[5025]: I1004 10:53:37.902309 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:38 crc kubenswrapper[5025]: W1004 10:53:38.356222 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70c3c42b_9625_43a5_9d21_a145d8380fc0.slice/crio-7d3557261df521eeb56917209311c8cedb2c4fd5e20ee4a9e744f4f37b917180 WatchSource:0}: Error finding container 7d3557261df521eeb56917209311c8cedb2c4fd5e20ee4a9e744f4f37b917180: Status 404 returned error can't find the container with id 7d3557261df521eeb56917209311c8cedb2c4fd5e20ee4a9e744f4f37b917180 Oct 04 10:53:38 crc kubenswrapper[5025]: I1004 10:53:38.356681 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fdcd7"] Oct 04 10:53:38 crc kubenswrapper[5025]: I1004 10:53:38.837290 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2aee4089-d0b3-4caf-89a6-70902456cf61","Type":"ContainerStarted","Data":"e7aecdfe34a0d63d9be74d67941096c09da16a882e478d841f78b34c14d77b69"} Oct 04 10:53:38 crc kubenswrapper[5025]: I1004 10:53:38.838125 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fdcd7" event={"ID":"70c3c42b-9625-43a5-9d21-a145d8380fc0","Type":"ContainerStarted","Data":"7d3557261df521eeb56917209311c8cedb2c4fd5e20ee4a9e744f4f37b917180"} Oct 04 10:53:39 crc kubenswrapper[5025]: I1004 10:53:39.851944 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2aee4089-d0b3-4caf-89a6-70902456cf61","Type":"ContainerStarted","Data":"c60174273b9368499dfd1949482ff92944d69ce229b49866aebc9fb249f04a97"} Oct 04 10:53:40 crc kubenswrapper[5025]: I1004 10:53:40.870322 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2aee4089-d0b3-4caf-89a6-70902456cf61","Type":"ContainerStarted","Data":"cdac6c9b11e5c5859f6d9ef8369e4929957ca2d5465d0fb013a5328ab760cfba"} Oct 04 10:53:43 crc kubenswrapper[5025]: I1004 10:53:43.379988 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 10:53:43 crc kubenswrapper[5025]: I1004 10:53:43.380357 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 04 10:53:43 crc kubenswrapper[5025]: I1004 10:53:43.422758 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 10:53:43 crc kubenswrapper[5025]: I1004 10:53:43.428392 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 04 10:53:43 crc kubenswrapper[5025]: I1004 10:53:43.896612 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 10:53:43 crc kubenswrapper[5025]: I1004 10:53:43.896662 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 04 10:53:44 crc kubenswrapper[5025]: I1004 10:53:44.098614 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:44 crc kubenswrapper[5025]: I1004 10:53:44.099626 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:44 crc kubenswrapper[5025]: I1004 10:53:44.156498 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:44 crc kubenswrapper[5025]: I1004 10:53:44.159981 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:44 crc kubenswrapper[5025]: I1004 10:53:44.907992 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:44 crc kubenswrapper[5025]: I1004 10:53:44.908090 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:45 crc kubenswrapper[5025]: I1004 10:53:45.752272 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 10:53:45 crc kubenswrapper[5025]: I1004 10:53:45.803440 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 04 10:53:46 crc kubenswrapper[5025]: I1004 10:53:46.913492 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:46 crc kubenswrapper[5025]: I1004 10:53:46.930261 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fdcd7" event={"ID":"70c3c42b-9625-43a5-9d21-a145d8380fc0","Type":"ContainerStarted","Data":"cc4e93f993fe836c8759233ae9ab35fee56e949b8b2916fa754a0f4196d7780d"} Oct 04 10:53:46 crc kubenswrapper[5025]: I1004 10:53:46.935118 5025 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 04 10:53:46 crc kubenswrapper[5025]: I1004 10:53:46.935186 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2aee4089-d0b3-4caf-89a6-70902456cf61","Type":"ContainerStarted","Data":"434a7b0d6c6fb2f65ab8f3896c1218e304db26b0a9d3e7f8eca462fe8b0d5e38"} Oct 04 10:53:46 crc kubenswrapper[5025]: I1004 10:53:46.936347 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 10:53:46 crc kubenswrapper[5025]: I1004 10:53:46.952659 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-fdcd7" podStartSLOduration=2.060672091 podStartE2EDuration="9.952641309s" podCreationTimestamp="2025-10-04 10:53:37 +0000 UTC" firstStartedPulling="2025-10-04 10:53:38.357773807 +0000 UTC m=+1146.782740687" lastFinishedPulling="2025-10-04 10:53:46.249743025 +0000 UTC m=+1154.674709905" observedRunningTime="2025-10-04 10:53:46.950446388 +0000 UTC m=+1155.375413268" watchObservedRunningTime="2025-10-04 10:53:46.952641309 +0000 UTC m=+1155.377608199" Oct 04 10:53:46 crc kubenswrapper[5025]: I1004 10:53:46.975461 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.440570108 podStartE2EDuration="11.975419229s" podCreationTimestamp="2025-10-04 10:53:35 +0000 UTC" firstStartedPulling="2025-10-04 10:53:36.71727414 +0000 UTC m=+1145.142241020" lastFinishedPulling="2025-10-04 10:53:46.252123261 +0000 UTC m=+1154.677090141" observedRunningTime="2025-10-04 10:53:46.971524061 +0000 UTC m=+1155.396491111" watchObservedRunningTime="2025-10-04 10:53:46.975419229 +0000 UTC m=+1155.400386109" Oct 04 10:53:47 crc kubenswrapper[5025]: I1004 10:53:47.026735 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 04 10:53:57 crc kubenswrapper[5025]: I1004 10:53:57.048562 5025 generic.go:334] "Generic (PLEG): container finished" podID="70c3c42b-9625-43a5-9d21-a145d8380fc0" containerID="cc4e93f993fe836c8759233ae9ab35fee56e949b8b2916fa754a0f4196d7780d" exitCode=0 Oct 04 10:53:57 crc kubenswrapper[5025]: I1004 10:53:57.049000 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fdcd7" event={"ID":"70c3c42b-9625-43a5-9d21-a145d8380fc0","Type":"ContainerDied","Data":"cc4e93f993fe836c8759233ae9ab35fee56e949b8b2916fa754a0f4196d7780d"} Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.417026 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.516529 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd2ds\" (UniqueName: \"kubernetes.io/projected/70c3c42b-9625-43a5-9d21-a145d8380fc0-kube-api-access-wd2ds\") pod \"70c3c42b-9625-43a5-9d21-a145d8380fc0\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.516677 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-combined-ca-bundle\") pod \"70c3c42b-9625-43a5-9d21-a145d8380fc0\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.516784 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-config-data\") pod \"70c3c42b-9625-43a5-9d21-a145d8380fc0\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.516966 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-scripts\") pod \"70c3c42b-9625-43a5-9d21-a145d8380fc0\" (UID: \"70c3c42b-9625-43a5-9d21-a145d8380fc0\") " Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.523312 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-scripts" (OuterVolumeSpecName: "scripts") pod "70c3c42b-9625-43a5-9d21-a145d8380fc0" (UID: "70c3c42b-9625-43a5-9d21-a145d8380fc0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.540402 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70c3c42b-9625-43a5-9d21-a145d8380fc0-kube-api-access-wd2ds" (OuterVolumeSpecName: "kube-api-access-wd2ds") pod "70c3c42b-9625-43a5-9d21-a145d8380fc0" (UID: "70c3c42b-9625-43a5-9d21-a145d8380fc0"). InnerVolumeSpecName "kube-api-access-wd2ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.543980 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-config-data" (OuterVolumeSpecName: "config-data") pod "70c3c42b-9625-43a5-9d21-a145d8380fc0" (UID: "70c3c42b-9625-43a5-9d21-a145d8380fc0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.551716 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70c3c42b-9625-43a5-9d21-a145d8380fc0" (UID: "70c3c42b-9625-43a5-9d21-a145d8380fc0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.620766 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd2ds\" (UniqueName: \"kubernetes.io/projected/70c3c42b-9625-43a5-9d21-a145d8380fc0-kube-api-access-wd2ds\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.620803 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.620812 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:58 crc kubenswrapper[5025]: I1004 10:53:58.620820 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70c3c42b-9625-43a5-9d21-a145d8380fc0-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.074768 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fdcd7" event={"ID":"70c3c42b-9625-43a5-9d21-a145d8380fc0","Type":"ContainerDied","Data":"7d3557261df521eeb56917209311c8cedb2c4fd5e20ee4a9e744f4f37b917180"} Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.075123 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d3557261df521eeb56917209311c8cedb2c4fd5e20ee4a9e744f4f37b917180" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.075356 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fdcd7" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.270842 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 10:53:59 crc kubenswrapper[5025]: E1004 10:53:59.271341 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70c3c42b-9625-43a5-9d21-a145d8380fc0" containerName="nova-cell0-conductor-db-sync" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.271362 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="70c3c42b-9625-43a5-9d21-a145d8380fc0" containerName="nova-cell0-conductor-db-sync" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.271583 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="70c3c42b-9625-43a5-9d21-a145d8380fc0" containerName="nova-cell0-conductor-db-sync" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.272345 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.274845 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.283643 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vxxmc" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.295319 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.336054 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cfb2082-5742-4e6f-aedf-a6e22e661ad3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3cfb2082-5742-4e6f-aedf-a6e22e661ad3\") " pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.336155 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cfb2082-5742-4e6f-aedf-a6e22e661ad3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3cfb2082-5742-4e6f-aedf-a6e22e661ad3\") " pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.336306 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkkhh\" (UniqueName: \"kubernetes.io/projected/3cfb2082-5742-4e6f-aedf-a6e22e661ad3-kube-api-access-hkkhh\") pod \"nova-cell0-conductor-0\" (UID: \"3cfb2082-5742-4e6f-aedf-a6e22e661ad3\") " pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.438151 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkkhh\" (UniqueName: \"kubernetes.io/projected/3cfb2082-5742-4e6f-aedf-a6e22e661ad3-kube-api-access-hkkhh\") pod \"nova-cell0-conductor-0\" (UID: \"3cfb2082-5742-4e6f-aedf-a6e22e661ad3\") " pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.438243 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cfb2082-5742-4e6f-aedf-a6e22e661ad3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3cfb2082-5742-4e6f-aedf-a6e22e661ad3\") " pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.438326 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cfb2082-5742-4e6f-aedf-a6e22e661ad3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3cfb2082-5742-4e6f-aedf-a6e22e661ad3\") " pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.443906 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cfb2082-5742-4e6f-aedf-a6e22e661ad3-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3cfb2082-5742-4e6f-aedf-a6e22e661ad3\") " pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.445595 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cfb2082-5742-4e6f-aedf-a6e22e661ad3-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3cfb2082-5742-4e6f-aedf-a6e22e661ad3\") " pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.459809 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkkhh\" (UniqueName: \"kubernetes.io/projected/3cfb2082-5742-4e6f-aedf-a6e22e661ad3-kube-api-access-hkkhh\") pod \"nova-cell0-conductor-0\" (UID: \"3cfb2082-5742-4e6f-aedf-a6e22e661ad3\") " pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.594413 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 04 10:53:59 crc kubenswrapper[5025]: I1004 10:53:59.997166 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 04 10:54:00 crc kubenswrapper[5025]: I1004 10:54:00.085276 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3cfb2082-5742-4e6f-aedf-a6e22e661ad3","Type":"ContainerStarted","Data":"62105bb1c5fe5ada81c7f6e768ee96fd2bfcec393e875e8a11b1542ca60353cc"} Oct 04 10:54:01 crc kubenswrapper[5025]: I1004 10:54:01.096103 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3cfb2082-5742-4e6f-aedf-a6e22e661ad3","Type":"ContainerStarted","Data":"142be6fffec0bc7fced8a487d72359b3ba4c481637daec78ce36c23e0f4b3520"} Oct 04 10:54:01 crc kubenswrapper[5025]: I1004 10:54:01.096634 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 04 10:54:01 crc kubenswrapper[5025]: I1004 10:54:01.117224 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.117208555 podStartE2EDuration="2.117208555s" podCreationTimestamp="2025-10-04 10:53:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:01.115259881 +0000 UTC m=+1169.540226801" watchObservedRunningTime="2025-10-04 10:54:01.117208555 +0000 UTC m=+1169.542175435" Oct 04 10:54:06 crc kubenswrapper[5025]: I1004 10:54:06.231279 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 10:54:09 crc kubenswrapper[5025]: I1004 10:54:09.634079 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.230403 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-tpdlf"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.232828 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.236840 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.237186 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.248273 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-tpdlf"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.273395 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5qt2\" (UniqueName: \"kubernetes.io/projected/03d77cd8-31b2-4500-92c3-55ca3c0d09de-kube-api-access-b5qt2\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.273958 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-scripts\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.274211 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-config-data\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.274330 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.371805 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.373515 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.375725 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-config-data\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.375780 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.375856 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5qt2\" (UniqueName: \"kubernetes.io/projected/03d77cd8-31b2-4500-92c3-55ca3c0d09de-kube-api-access-b5qt2\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.375933 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs4df\" (UniqueName: \"kubernetes.io/projected/37905f14-a616-4f33-854d-e7313dd23b5d-kube-api-access-rs4df\") pod \"nova-scheduler-0\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.375962 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.376078 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-config-data\") pod \"nova-scheduler-0\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.376109 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-scripts\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.379871 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.396732 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.399526 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-scripts\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.417094 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-config-data\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.419580 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5qt2\" (UniqueName: \"kubernetes.io/projected/03d77cd8-31b2-4500-92c3-55ca3c0d09de-kube-api-access-b5qt2\") pod \"nova-cell0-cell-mapping-tpdlf\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.462338 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.500554 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs4df\" (UniqueName: \"kubernetes.io/projected/37905f14-a616-4f33-854d-e7313dd23b5d-kube-api-access-rs4df\") pod \"nova-scheduler-0\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.500629 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.500776 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-config-data\") pod \"nova-scheduler-0\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.514172 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-config-data\") pod \"nova-scheduler-0\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.515434 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.563810 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.565898 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs4df\" (UniqueName: \"kubernetes.io/projected/37905f14-a616-4f33-854d-e7313dd23b5d-kube-api-access-rs4df\") pod \"nova-scheduler-0\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.566430 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.566566 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.579467 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.585103 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.600265 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.601872 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.603530 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.603709 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.625274 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-config-data\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.625322 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bzhw\" (UniqueName: \"kubernetes.io/projected/2e6a4f0f-8574-434d-93d0-79a40c76776a-kube-api-access-9bzhw\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.625434 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5btbk\" (UniqueName: \"kubernetes.io/projected/ec9f4766-d4bf-4db8-851e-368d670a5fa7-kube-api-access-5btbk\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.625459 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.625483 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec9f4766-d4bf-4db8-851e-368d670a5fa7-logs\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.625509 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e6a4f0f-8574-434d-93d0-79a40c76776a-logs\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.625548 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-config-data\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.625580 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.635313 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.635495 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="8dbf11e2-e00c-4f47-b1cf-07223dac4842" containerName="kube-state-metrics" containerID="cri-o://a0859cf37a7b14d6a0f6a57b9e16088a969791aa10112e37ce5663c76ab17ed5" gracePeriod=30 Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.668734 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.727255 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5btbk\" (UniqueName: \"kubernetes.io/projected/ec9f4766-d4bf-4db8-851e-368d670a5fa7-kube-api-access-5btbk\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.727302 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.727334 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec9f4766-d4bf-4db8-851e-368d670a5fa7-logs\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.727360 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e6a4f0f-8574-434d-93d0-79a40c76776a-logs\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.727420 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-config-data\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.727455 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.727475 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-config-data\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.727493 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bzhw\" (UniqueName: \"kubernetes.io/projected/2e6a4f0f-8574-434d-93d0-79a40c76776a-kube-api-access-9bzhw\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.728251 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec9f4766-d4bf-4db8-851e-368d670a5fa7-logs\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.728529 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e6a4f0f-8574-434d-93d0-79a40c76776a-logs\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.734906 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.737551 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.744532 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-ngr4b"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.746458 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.767604 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bzhw\" (UniqueName: \"kubernetes.io/projected/2e6a4f0f-8574-434d-93d0-79a40c76776a-kube-api-access-9bzhw\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.782918 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5btbk\" (UniqueName: \"kubernetes.io/projected/ec9f4766-d4bf-4db8-851e-368d670a5fa7-kube-api-access-5btbk\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.790664 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-config-data\") pod \"nova-api-0\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " pod="openstack/nova-api-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.791283 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-config-data\") pod \"nova-metadata-0\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " pod="openstack/nova-metadata-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.807698 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-ngr4b"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.834229 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-svc\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.834391 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-config\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.834447 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.834485 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.834523 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.834562 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2czh\" (UniqueName: \"kubernetes.io/projected/6d8b4798-48b3-40d0-8330-2c7294991b3e-kube-api-access-n2czh\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.935779 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2czh\" (UniqueName: \"kubernetes.io/projected/6d8b4798-48b3-40d0-8330-2c7294991b3e-kube-api-access-n2czh\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.936275 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-svc\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.936333 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-config\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.936361 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.936389 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.936404 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.937991 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-svc\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.938562 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.939301 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-config\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.939819 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.939969 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.940845 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.942174 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.947777 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.963697 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 10:54:10 crc kubenswrapper[5025]: I1004 10:54:10.978870 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2czh\" (UniqueName: \"kubernetes.io/projected/6d8b4798-48b3-40d0-8330-2c7294991b3e-kube-api-access-n2czh\") pod \"dnsmasq-dns-865f5d856f-ngr4b\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.028589 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.038562 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6qvq\" (UniqueName: \"kubernetes.io/projected/dc28e385-af83-45f7-9bb3-25edf85e2866-kube-api-access-w6qvq\") pod \"nova-cell1-novncproxy-0\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.038672 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.038710 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.058852 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.118909 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.140247 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6qvq\" (UniqueName: \"kubernetes.io/projected/dc28e385-af83-45f7-9bb3-25edf85e2866-kube-api-access-w6qvq\") pod \"nova-cell1-novncproxy-0\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.140753 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.140818 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.147725 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.149984 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.162510 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6qvq\" (UniqueName: \"kubernetes.io/projected/dc28e385-af83-45f7-9bb3-25edf85e2866-kube-api-access-w6qvq\") pod \"nova-cell1-novncproxy-0\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.245507 5025 generic.go:334] "Generic (PLEG): container finished" podID="8dbf11e2-e00c-4f47-b1cf-07223dac4842" containerID="a0859cf37a7b14d6a0f6a57b9e16088a969791aa10112e37ce5663c76ab17ed5" exitCode=2 Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.245567 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8dbf11e2-e00c-4f47-b1cf-07223dac4842","Type":"ContainerDied","Data":"a0859cf37a7b14d6a0f6a57b9e16088a969791aa10112e37ce5663c76ab17ed5"} Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.301238 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.497271 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-tpdlf"] Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.601440 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.714717 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.760165 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wdjvg"] Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.761290 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.768228 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.768424 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.770053 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wdjvg"] Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.844258 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-ngr4b"] Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.854548 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.854627 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-scripts\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.854735 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9ksn\" (UniqueName: \"kubernetes.io/projected/57e9191a-aa60-498f-b621-2b83a7b795ed-kube-api-access-x9ksn\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.854917 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-config-data\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.877051 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.893102 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.968964 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxhzh\" (UniqueName: \"kubernetes.io/projected/8dbf11e2-e00c-4f47-b1cf-07223dac4842-kube-api-access-cxhzh\") pod \"8dbf11e2-e00c-4f47-b1cf-07223dac4842\" (UID: \"8dbf11e2-e00c-4f47-b1cf-07223dac4842\") " Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.969328 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9ksn\" (UniqueName: \"kubernetes.io/projected/57e9191a-aa60-498f-b621-2b83a7b795ed-kube-api-access-x9ksn\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.969489 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-config-data\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.969582 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.969638 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-scripts\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.980759 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-config-data\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.980794 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dbf11e2-e00c-4f47-b1cf-07223dac4842-kube-api-access-cxhzh" (OuterVolumeSpecName: "kube-api-access-cxhzh") pod "8dbf11e2-e00c-4f47-b1cf-07223dac4842" (UID: "8dbf11e2-e00c-4f47-b1cf-07223dac4842"). InnerVolumeSpecName "kube-api-access-cxhzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.982920 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-scripts\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.983037 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:11 crc kubenswrapper[5025]: I1004 10:54:11.988905 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9ksn\" (UniqueName: \"kubernetes.io/projected/57e9191a-aa60-498f-b621-2b83a7b795ed-kube-api-access-x9ksn\") pod \"nova-cell1-conductor-db-sync-wdjvg\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.057700 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.071129 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxhzh\" (UniqueName: \"kubernetes.io/projected/8dbf11e2-e00c-4f47-b1cf-07223dac4842-kube-api-access-cxhzh\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.108203 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.264896 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dc28e385-af83-45f7-9bb3-25edf85e2866","Type":"ContainerStarted","Data":"06d959aef2a731c121680f729a7e5f1c75879fe86ac5b9de722554fd8805cd53"} Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.275538 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8dbf11e2-e00c-4f47-b1cf-07223dac4842","Type":"ContainerDied","Data":"a825b59431aeff480503a24f3513b7a36c75bb82611afaf11af14fba997a4dda"} Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.275598 5025 scope.go:117] "RemoveContainer" containerID="a0859cf37a7b14d6a0f6a57b9e16088a969791aa10112e37ce5663c76ab17ed5" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.275762 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.287969 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec9f4766-d4bf-4db8-851e-368d670a5fa7","Type":"ContainerStarted","Data":"a4fae44b981a747f25944aec6124810595ab50d0b3bd62e5b1acdc7c1bffe7e9"} Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.296727 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tpdlf" event={"ID":"03d77cd8-31b2-4500-92c3-55ca3c0d09de","Type":"ContainerStarted","Data":"ce3d995cbaac27290050e3da1c38213b610291d8960d46702cf5decf6093bd1b"} Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.296774 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tpdlf" event={"ID":"03d77cd8-31b2-4500-92c3-55ca3c0d09de","Type":"ContainerStarted","Data":"e1e1020725fa1e5fb6d972b198c85497b37071ea30551b899c106fb157defdb4"} Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.331076 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37905f14-a616-4f33-854d-e7313dd23b5d","Type":"ContainerStarted","Data":"18c3c54a566058a9b103dc5f385857554677d30b69777c2fbcdaf1cfa50eb211"} Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.339229 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.340127 5025 generic.go:334] "Generic (PLEG): container finished" podID="6d8b4798-48b3-40d0-8330-2c7294991b3e" containerID="40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09" exitCode=0 Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.341281 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" event={"ID":"6d8b4798-48b3-40d0-8330-2c7294991b3e","Type":"ContainerDied","Data":"40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09"} Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.341325 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" event={"ID":"6d8b4798-48b3-40d0-8330-2c7294991b3e","Type":"ContainerStarted","Data":"7abad6fce4504da7ec274c39bf669ed2e8be13d2725ea92e1279467096df2214"} Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.359274 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e6a4f0f-8574-434d-93d0-79a40c76776a","Type":"ContainerStarted","Data":"fc0c66196791fb43ed25ef9a13321f53c3f55f744d71c09394f1f7b3b6ebbf49"} Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.360448 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.398193 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 10:54:12 crc kubenswrapper[5025]: E1004 10:54:12.398605 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dbf11e2-e00c-4f47-b1cf-07223dac4842" containerName="kube-state-metrics" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.398617 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dbf11e2-e00c-4f47-b1cf-07223dac4842" containerName="kube-state-metrics" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.398776 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dbf11e2-e00c-4f47-b1cf-07223dac4842" containerName="kube-state-metrics" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.399409 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.401562 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.401733 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.434774 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dbf11e2-e00c-4f47-b1cf-07223dac4842" path="/var/lib/kubelet/pods/8dbf11e2-e00c-4f47-b1cf-07223dac4842/volumes" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.435731 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.443529 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-tpdlf" podStartSLOduration=2.443504132 podStartE2EDuration="2.443504132s" podCreationTimestamp="2025-10-04 10:54:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:12.359320916 +0000 UTC m=+1180.784287796" watchObservedRunningTime="2025-10-04 10:54:12.443504132 +0000 UTC m=+1180.868471012" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.486154 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/16cbe27e-1c01-4140-abbb-351ddb59d907-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.486213 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/16cbe27e-1c01-4140-abbb-351ddb59d907-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.503561 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16cbe27e-1c01-4140-abbb-351ddb59d907-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.503642 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49nzd\" (UniqueName: \"kubernetes.io/projected/16cbe27e-1c01-4140-abbb-351ddb59d907-kube-api-access-49nzd\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.609432 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/16cbe27e-1c01-4140-abbb-351ddb59d907-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.610003 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16cbe27e-1c01-4140-abbb-351ddb59d907-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.610059 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49nzd\" (UniqueName: \"kubernetes.io/projected/16cbe27e-1c01-4140-abbb-351ddb59d907-kube-api-access-49nzd\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.610101 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/16cbe27e-1c01-4140-abbb-351ddb59d907-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.615869 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/16cbe27e-1c01-4140-abbb-351ddb59d907-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.617045 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16cbe27e-1c01-4140-abbb-351ddb59d907-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.620230 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/16cbe27e-1c01-4140-abbb-351ddb59d907-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.627996 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49nzd\" (UniqueName: \"kubernetes.io/projected/16cbe27e-1c01-4140-abbb-351ddb59d907-kube-api-access-49nzd\") pod \"kube-state-metrics-0\" (UID: \"16cbe27e-1c01-4140-abbb-351ddb59d907\") " pod="openstack/kube-state-metrics-0" Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.631487 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wdjvg"] Oct 04 10:54:12 crc kubenswrapper[5025]: I1004 10:54:12.838562 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.389966 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" event={"ID":"6d8b4798-48b3-40d0-8330-2c7294991b3e","Type":"ContainerStarted","Data":"fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35"} Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.390248 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.395410 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wdjvg" event={"ID":"57e9191a-aa60-498f-b621-2b83a7b795ed","Type":"ContainerStarted","Data":"10e5e89d8d185b91a6ff58dbe1360f1343045faa872a02142d18c5b1cb774ab6"} Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.395463 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wdjvg" event={"ID":"57e9191a-aa60-498f-b621-2b83a7b795ed","Type":"ContainerStarted","Data":"b0dafdbcd05230a453dd70f523b2192d0095ff1eb306414336b457a11e4822a0"} Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.421689 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" podStartSLOduration=3.421662534 podStartE2EDuration="3.421662534s" podCreationTimestamp="2025-10-04 10:54:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:13.4048507 +0000 UTC m=+1181.829817590" watchObservedRunningTime="2025-10-04 10:54:13.421662534 +0000 UTC m=+1181.846629414" Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.458930 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-wdjvg" podStartSLOduration=2.4588966230000002 podStartE2EDuration="2.458896623s" podCreationTimestamp="2025-10-04 10:54:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:13.42693499 +0000 UTC m=+1181.851901870" watchObservedRunningTime="2025-10-04 10:54:13.458896623 +0000 UTC m=+1181.883863503" Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.563574 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.955283 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.955548 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="ceilometer-central-agent" containerID="cri-o://e7aecdfe34a0d63d9be74d67941096c09da16a882e478d841f78b34c14d77b69" gracePeriod=30 Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.955920 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="proxy-httpd" containerID="cri-o://434a7b0d6c6fb2f65ab8f3896c1218e304db26b0a9d3e7f8eca462fe8b0d5e38" gracePeriod=30 Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.955967 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="sg-core" containerID="cri-o://cdac6c9b11e5c5859f6d9ef8369e4929957ca2d5465d0fb013a5328ab760cfba" gracePeriod=30 Oct 04 10:54:13 crc kubenswrapper[5025]: I1004 10:54:13.956342 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="ceilometer-notification-agent" containerID="cri-o://c60174273b9368499dfd1949482ff92944d69ce229b49866aebc9fb249f04a97" gracePeriod=30 Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.407315 5025 generic.go:334] "Generic (PLEG): container finished" podID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerID="434a7b0d6c6fb2f65ab8f3896c1218e304db26b0a9d3e7f8eca462fe8b0d5e38" exitCode=0 Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.407561 5025 generic.go:334] "Generic (PLEG): container finished" podID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerID="cdac6c9b11e5c5859f6d9ef8369e4929957ca2d5465d0fb013a5328ab760cfba" exitCode=2 Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.407570 5025 generic.go:334] "Generic (PLEG): container finished" podID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerID="e7aecdfe34a0d63d9be74d67941096c09da16a882e478d841f78b34c14d77b69" exitCode=0 Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.407381 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2aee4089-d0b3-4caf-89a6-70902456cf61","Type":"ContainerDied","Data":"434a7b0d6c6fb2f65ab8f3896c1218e304db26b0a9d3e7f8eca462fe8b0d5e38"} Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.407625 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2aee4089-d0b3-4caf-89a6-70902456cf61","Type":"ContainerDied","Data":"cdac6c9b11e5c5859f6d9ef8369e4929957ca2d5465d0fb013a5328ab760cfba"} Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.407635 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2aee4089-d0b3-4caf-89a6-70902456cf61","Type":"ContainerDied","Data":"e7aecdfe34a0d63d9be74d67941096c09da16a882e478d841f78b34c14d77b69"} Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.410121 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"16cbe27e-1c01-4140-abbb-351ddb59d907","Type":"ContainerStarted","Data":"3cc713eb06c5591062bce45d863741e611750f8004cd4837e76ae311f48a6ed1"} Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.654926 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.669035 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.714203 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:54:14 crc kubenswrapper[5025]: I1004 10:54:14.714262 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:54:15 crc kubenswrapper[5025]: I1004 10:54:15.420767 5025 generic.go:334] "Generic (PLEG): container finished" podID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerID="c60174273b9368499dfd1949482ff92944d69ce229b49866aebc9fb249f04a97" exitCode=0 Oct 04 10:54:15 crc kubenswrapper[5025]: I1004 10:54:15.420816 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2aee4089-d0b3-4caf-89a6-70902456cf61","Type":"ContainerDied","Data":"c60174273b9368499dfd1949482ff92944d69ce229b49866aebc9fb249f04a97"} Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.230944 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.293230 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-sg-core-conf-yaml\") pod \"2aee4089-d0b3-4caf-89a6-70902456cf61\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.293270 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-run-httpd\") pod \"2aee4089-d0b3-4caf-89a6-70902456cf61\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.293363 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-log-httpd\") pod \"2aee4089-d0b3-4caf-89a6-70902456cf61\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.293389 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-combined-ca-bundle\") pod \"2aee4089-d0b3-4caf-89a6-70902456cf61\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.293408 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg2c2\" (UniqueName: \"kubernetes.io/projected/2aee4089-d0b3-4caf-89a6-70902456cf61-kube-api-access-pg2c2\") pod \"2aee4089-d0b3-4caf-89a6-70902456cf61\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.293508 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-scripts\") pod \"2aee4089-d0b3-4caf-89a6-70902456cf61\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.293553 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-config-data\") pod \"2aee4089-d0b3-4caf-89a6-70902456cf61\" (UID: \"2aee4089-d0b3-4caf-89a6-70902456cf61\") " Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.293717 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2aee4089-d0b3-4caf-89a6-70902456cf61" (UID: "2aee4089-d0b3-4caf-89a6-70902456cf61"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.293913 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2aee4089-d0b3-4caf-89a6-70902456cf61" (UID: "2aee4089-d0b3-4caf-89a6-70902456cf61"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.293978 5025 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.304588 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-scripts" (OuterVolumeSpecName: "scripts") pod "2aee4089-d0b3-4caf-89a6-70902456cf61" (UID: "2aee4089-d0b3-4caf-89a6-70902456cf61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.316140 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aee4089-d0b3-4caf-89a6-70902456cf61-kube-api-access-pg2c2" (OuterVolumeSpecName: "kube-api-access-pg2c2") pod "2aee4089-d0b3-4caf-89a6-70902456cf61" (UID: "2aee4089-d0b3-4caf-89a6-70902456cf61"). InnerVolumeSpecName "kube-api-access-pg2c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.395435 5025 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2aee4089-d0b3-4caf-89a6-70902456cf61-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.395476 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg2c2\" (UniqueName: \"kubernetes.io/projected/2aee4089-d0b3-4caf-89a6-70902456cf61-kube-api-access-pg2c2\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.395486 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.451195 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2aee4089-d0b3-4caf-89a6-70902456cf61" (UID: "2aee4089-d0b3-4caf-89a6-70902456cf61"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.476783 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec9f4766-d4bf-4db8-851e-368d670a5fa7","Type":"ContainerStarted","Data":"08c5ac95c8a494b2c84c8f29df3f940bd79ee23fee759ecba9b6c7c8ac4066d7"} Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.476824 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec9f4766-d4bf-4db8-851e-368d670a5fa7","Type":"ContainerStarted","Data":"6c76fe1b0b99182de30bf2bae872be46b21b7795e4170461121f739587d9b833"} Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.476946 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" containerName="nova-metadata-log" containerID="cri-o://6c76fe1b0b99182de30bf2bae872be46b21b7795e4170461121f739587d9b833" gracePeriod=30 Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.477433 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" containerName="nova-metadata-metadata" containerID="cri-o://08c5ac95c8a494b2c84c8f29df3f940bd79ee23fee759ecba9b6c7c8ac4066d7" gracePeriod=30 Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.496479 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37905f14-a616-4f33-854d-e7313dd23b5d","Type":"ContainerStarted","Data":"a245353f4935b5ca0f2970b2560b03ee4e2a18645bd6a59025e321ee7263dc7b"} Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.497905 5025 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.505365 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.522637961 podStartE2EDuration="6.504527591s" podCreationTimestamp="2025-10-04 10:54:10 +0000 UTC" firstStartedPulling="2025-10-04 10:54:11.797300685 +0000 UTC m=+1180.222267565" lastFinishedPulling="2025-10-04 10:54:15.779190325 +0000 UTC m=+1184.204157195" observedRunningTime="2025-10-04 10:54:16.496849259 +0000 UTC m=+1184.921816139" watchObservedRunningTime="2025-10-04 10:54:16.504527591 +0000 UTC m=+1184.929494471" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.514174 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2aee4089-d0b3-4caf-89a6-70902456cf61" (UID: "2aee4089-d0b3-4caf-89a6-70902456cf61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.527100 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.434165505 podStartE2EDuration="6.527078094s" podCreationTimestamp="2025-10-04 10:54:10 +0000 UTC" firstStartedPulling="2025-10-04 10:54:11.643324499 +0000 UTC m=+1180.068291369" lastFinishedPulling="2025-10-04 10:54:15.736237068 +0000 UTC m=+1184.161203958" observedRunningTime="2025-10-04 10:54:16.521821079 +0000 UTC m=+1184.946787979" watchObservedRunningTime="2025-10-04 10:54:16.527078094 +0000 UTC m=+1184.952044974" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.531688 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e6a4f0f-8574-434d-93d0-79a40c76776a","Type":"ContainerStarted","Data":"a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d"} Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.531733 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e6a4f0f-8574-434d-93d0-79a40c76776a","Type":"ContainerStarted","Data":"bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f"} Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.547488 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-config-data" (OuterVolumeSpecName: "config-data") pod "2aee4089-d0b3-4caf-89a6-70902456cf61" (UID: "2aee4089-d0b3-4caf-89a6-70902456cf61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.551390 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2aee4089-d0b3-4caf-89a6-70902456cf61","Type":"ContainerDied","Data":"42ea50d7f3f4a4ca840aa0b40f62b04fad46a4e766c25c2b808169a9916fd480"} Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.551437 5025 scope.go:117] "RemoveContainer" containerID="434a7b0d6c6fb2f65ab8f3896c1218e304db26b0a9d3e7f8eca462fe8b0d5e38" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.551582 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.559802 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.680792121 podStartE2EDuration="6.559782888s" podCreationTimestamp="2025-10-04 10:54:10 +0000 UTC" firstStartedPulling="2025-10-04 10:54:11.858105195 +0000 UTC m=+1180.283072075" lastFinishedPulling="2025-10-04 10:54:15.737095962 +0000 UTC m=+1184.162062842" observedRunningTime="2025-10-04 10:54:16.556484857 +0000 UTC m=+1184.981451737" watchObservedRunningTime="2025-10-04 10:54:16.559782888 +0000 UTC m=+1184.984749768" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.565525 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dc28e385-af83-45f7-9bb3-25edf85e2866","Type":"ContainerStarted","Data":"14743e2481323cb7ef3ffe0099e6662bca90c3c14cca7c7625b0dbd06bbaa8f6"} Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.565587 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="dc28e385-af83-45f7-9bb3-25edf85e2866" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://14743e2481323cb7ef3ffe0099e6662bca90c3c14cca7c7625b0dbd06bbaa8f6" gracePeriod=30 Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.569101 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"16cbe27e-1c01-4140-abbb-351ddb59d907","Type":"ContainerStarted","Data":"496f7f44a6b68afda616c3b27567c1ea7ea2f3228034666aa93117ef4e71eb75"} Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.569221 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.592063 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.890761882 podStartE2EDuration="6.592043109s" podCreationTimestamp="2025-10-04 10:54:10 +0000 UTC" firstStartedPulling="2025-10-04 10:54:12.075538743 +0000 UTC m=+1180.500505623" lastFinishedPulling="2025-10-04 10:54:15.77681996 +0000 UTC m=+1184.201786850" observedRunningTime="2025-10-04 10:54:16.588935823 +0000 UTC m=+1185.013902703" watchObservedRunningTime="2025-10-04 10:54:16.592043109 +0000 UTC m=+1185.017009989" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.599862 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.599891 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2aee4089-d0b3-4caf-89a6-70902456cf61-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.600173 5025 scope.go:117] "RemoveContainer" containerID="cdac6c9b11e5c5859f6d9ef8369e4929957ca2d5465d0fb013a5328ab760cfba" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.612534 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.464182335 podStartE2EDuration="4.612515365s" podCreationTimestamp="2025-10-04 10:54:12 +0000 UTC" firstStartedPulling="2025-10-04 10:54:13.580984047 +0000 UTC m=+1182.005950927" lastFinishedPulling="2025-10-04 10:54:15.729317077 +0000 UTC m=+1184.154283957" observedRunningTime="2025-10-04 10:54:16.608293278 +0000 UTC m=+1185.033260158" watchObservedRunningTime="2025-10-04 10:54:16.612515365 +0000 UTC m=+1185.037482245" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.628566 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.635180 5025 scope.go:117] "RemoveContainer" containerID="c60174273b9368499dfd1949482ff92944d69ce229b49866aebc9fb249f04a97" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.686967 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.697816 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:16 crc kubenswrapper[5025]: E1004 10:54:16.699003 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="proxy-httpd" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.699078 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="proxy-httpd" Oct 04 10:54:16 crc kubenswrapper[5025]: E1004 10:54:16.699143 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="sg-core" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.699152 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="sg-core" Oct 04 10:54:16 crc kubenswrapper[5025]: E1004 10:54:16.699201 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="ceilometer-notification-agent" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.699212 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="ceilometer-notification-agent" Oct 04 10:54:16 crc kubenswrapper[5025]: E1004 10:54:16.699232 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="ceilometer-central-agent" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.699240 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="ceilometer-central-agent" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.699852 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="sg-core" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.699874 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="ceilometer-notification-agent" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.699897 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="proxy-httpd" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.699917 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" containerName="ceilometer-central-agent" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.702805 5025 scope.go:117] "RemoveContainer" containerID="e7aecdfe34a0d63d9be74d67941096c09da16a882e478d841f78b34c14d77b69" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.703254 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.706542 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.707884 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.708621 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.721090 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.803084 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.803204 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-log-httpd\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.803263 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-scripts\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.803279 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-config-data\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.803318 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.803402 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdcsz\" (UniqueName: \"kubernetes.io/projected/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-kube-api-access-cdcsz\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.803442 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.803463 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-run-httpd\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.905453 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.905816 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-log-httpd\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.905847 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-scripts\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.905870 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-config-data\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.905899 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.906004 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdcsz\" (UniqueName: \"kubernetes.io/projected/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-kube-api-access-cdcsz\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.906097 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.906127 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-run-httpd\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.906504 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-run-httpd\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.908626 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-log-httpd\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.910980 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.911292 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.911431 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.912800 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-scripts\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.917187 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-config-data\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:16 crc kubenswrapper[5025]: I1004 10:54:16.926602 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdcsz\" (UniqueName: \"kubernetes.io/projected/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-kube-api-access-cdcsz\") pod \"ceilometer-0\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " pod="openstack/ceilometer-0" Oct 04 10:54:17 crc kubenswrapper[5025]: I1004 10:54:17.036897 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:54:17 crc kubenswrapper[5025]: I1004 10:54:17.546744 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:17 crc kubenswrapper[5025]: I1004 10:54:17.592690 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a","Type":"ContainerStarted","Data":"99939ceedec07f8253f6e6d51e4a5cdf63094384a4259d9f7d230ed34f3fe4e8"} Oct 04 10:54:17 crc kubenswrapper[5025]: I1004 10:54:17.594966 5025 generic.go:334] "Generic (PLEG): container finished" podID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" containerID="6c76fe1b0b99182de30bf2bae872be46b21b7795e4170461121f739587d9b833" exitCode=143 Oct 04 10:54:17 crc kubenswrapper[5025]: I1004 10:54:17.595095 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec9f4766-d4bf-4db8-851e-368d670a5fa7","Type":"ContainerDied","Data":"6c76fe1b0b99182de30bf2bae872be46b21b7795e4170461121f739587d9b833"} Oct 04 10:54:18 crc kubenswrapper[5025]: I1004 10:54:18.421313 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aee4089-d0b3-4caf-89a6-70902456cf61" path="/var/lib/kubelet/pods/2aee4089-d0b3-4caf-89a6-70902456cf61/volumes" Oct 04 10:54:18 crc kubenswrapper[5025]: I1004 10:54:18.605320 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a","Type":"ContainerStarted","Data":"6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076"} Oct 04 10:54:19 crc kubenswrapper[5025]: I1004 10:54:19.616671 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a","Type":"ContainerStarted","Data":"0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120"} Oct 04 10:54:20 crc kubenswrapper[5025]: I1004 10:54:20.628317 5025 generic.go:334] "Generic (PLEG): container finished" podID="03d77cd8-31b2-4500-92c3-55ca3c0d09de" containerID="ce3d995cbaac27290050e3da1c38213b610291d8960d46702cf5decf6093bd1b" exitCode=0 Oct 04 10:54:20 crc kubenswrapper[5025]: I1004 10:54:20.628376 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tpdlf" event={"ID":"03d77cd8-31b2-4500-92c3-55ca3c0d09de","Type":"ContainerDied","Data":"ce3d995cbaac27290050e3da1c38213b610291d8960d46702cf5decf6093bd1b"} Oct 04 10:54:20 crc kubenswrapper[5025]: I1004 10:54:20.633811 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a","Type":"ContainerStarted","Data":"714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073"} Oct 04 10:54:20 crc kubenswrapper[5025]: I1004 10:54:20.636376 5025 generic.go:334] "Generic (PLEG): container finished" podID="57e9191a-aa60-498f-b621-2b83a7b795ed" containerID="10e5e89d8d185b91a6ff58dbe1360f1343045faa872a02142d18c5b1cb774ab6" exitCode=0 Oct 04 10:54:20 crc kubenswrapper[5025]: I1004 10:54:20.636397 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wdjvg" event={"ID":"57e9191a-aa60-498f-b621-2b83a7b795ed","Type":"ContainerDied","Data":"10e5e89d8d185b91a6ff58dbe1360f1343045faa872a02142d18c5b1cb774ab6"} Oct 04 10:54:20 crc kubenswrapper[5025]: I1004 10:54:20.670251 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 10:54:20 crc kubenswrapper[5025]: I1004 10:54:20.670297 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 10:54:20 crc kubenswrapper[5025]: I1004 10:54:20.700604 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.029695 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.029771 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.060546 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.060601 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.121249 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.197476 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-89zdw"] Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.197716 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" podUID="a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" containerName="dnsmasq-dns" containerID="cri-o://3b78cd90ae0947943209e117223a1b25701dc2b8bd997f4eee6275518e3db983" gracePeriod=10 Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.301677 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.660269 5025 generic.go:334] "Generic (PLEG): container finished" podID="a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" containerID="3b78cd90ae0947943209e117223a1b25701dc2b8bd997f4eee6275518e3db983" exitCode=0 Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.660419 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" event={"ID":"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf","Type":"ContainerDied","Data":"3b78cd90ae0947943209e117223a1b25701dc2b8bd997f4eee6275518e3db983"} Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.667711 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a","Type":"ContainerStarted","Data":"4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0"} Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.703515 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.737031 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.748065 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.000486792 podStartE2EDuration="5.748022547s" podCreationTimestamp="2025-10-04 10:54:16 +0000 UTC" firstStartedPulling="2025-10-04 10:54:17.573600445 +0000 UTC m=+1185.998567325" lastFinishedPulling="2025-10-04 10:54:21.3211362 +0000 UTC m=+1189.746103080" observedRunningTime="2025-10-04 10:54:21.693942143 +0000 UTC m=+1190.118909023" watchObservedRunningTime="2025-10-04 10:54:21.748022547 +0000 UTC m=+1190.172989427" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.807659 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-swift-storage-0\") pod \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.808297 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-config\") pod \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.808330 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-svc\") pod \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.808452 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-nb\") pod \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.808477 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-sb\") pod \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.808495 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7q59\" (UniqueName: \"kubernetes.io/projected/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-kube-api-access-s7q59\") pod \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.825342 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-kube-api-access-s7q59" (OuterVolumeSpecName: "kube-api-access-s7q59") pod "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" (UID: "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf"). InnerVolumeSpecName "kube-api-access-s7q59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.890283 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" (UID: "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.907568 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-config" (OuterVolumeSpecName: "config") pod "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" (UID: "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.912686 5025 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.912715 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.912748 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7q59\" (UniqueName: \"kubernetes.io/projected/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-kube-api-access-s7q59\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.928500 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" (UID: "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:54:21 crc kubenswrapper[5025]: E1004 10:54:21.955289 5025 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-nb podName:a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf nodeName:}" failed. No retries permitted until 2025-10-04 10:54:22.455265315 +0000 UTC m=+1190.880232185 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-nb" (UniqueName: "kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-nb") pod "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" (UID: "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf") : error deleting /var/lib/kubelet/pods/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf/volume-subpaths: remove /var/lib/kubelet/pods/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf/volume-subpaths: no such file or directory Oct 04 10:54:21 crc kubenswrapper[5025]: I1004 10:54:21.955555 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" (UID: "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.017302 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.017337 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.134752 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.146704 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.147103 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.204299 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.220761 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5qt2\" (UniqueName: \"kubernetes.io/projected/03d77cd8-31b2-4500-92c3-55ca3c0d09de-kube-api-access-b5qt2\") pod \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.220930 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-combined-ca-bundle\") pod \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.220994 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-scripts\") pod \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.221145 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-config-data\") pod \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\" (UID: \"03d77cd8-31b2-4500-92c3-55ca3c0d09de\") " Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.224647 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-scripts" (OuterVolumeSpecName: "scripts") pod "03d77cd8-31b2-4500-92c3-55ca3c0d09de" (UID: "03d77cd8-31b2-4500-92c3-55ca3c0d09de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.232271 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d77cd8-31b2-4500-92c3-55ca3c0d09de-kube-api-access-b5qt2" (OuterVolumeSpecName: "kube-api-access-b5qt2") pod "03d77cd8-31b2-4500-92c3-55ca3c0d09de" (UID: "03d77cd8-31b2-4500-92c3-55ca3c0d09de"). InnerVolumeSpecName "kube-api-access-b5qt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.248556 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-config-data" (OuterVolumeSpecName: "config-data") pod "03d77cd8-31b2-4500-92c3-55ca3c0d09de" (UID: "03d77cd8-31b2-4500-92c3-55ca3c0d09de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.270501 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03d77cd8-31b2-4500-92c3-55ca3c0d09de" (UID: "03d77cd8-31b2-4500-92c3-55ca3c0d09de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.322969 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-config-data\") pod \"57e9191a-aa60-498f-b621-2b83a7b795ed\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.323052 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9ksn\" (UniqueName: \"kubernetes.io/projected/57e9191a-aa60-498f-b621-2b83a7b795ed-kube-api-access-x9ksn\") pod \"57e9191a-aa60-498f-b621-2b83a7b795ed\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.323092 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-combined-ca-bundle\") pod \"57e9191a-aa60-498f-b621-2b83a7b795ed\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.323108 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-scripts\") pod \"57e9191a-aa60-498f-b621-2b83a7b795ed\" (UID: \"57e9191a-aa60-498f-b621-2b83a7b795ed\") " Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.323536 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5qt2\" (UniqueName: \"kubernetes.io/projected/03d77cd8-31b2-4500-92c3-55ca3c0d09de-kube-api-access-b5qt2\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.323552 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.323561 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.323569 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03d77cd8-31b2-4500-92c3-55ca3c0d09de-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.329530 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e9191a-aa60-498f-b621-2b83a7b795ed-kube-api-access-x9ksn" (OuterVolumeSpecName: "kube-api-access-x9ksn") pod "57e9191a-aa60-498f-b621-2b83a7b795ed" (UID: "57e9191a-aa60-498f-b621-2b83a7b795ed"). InnerVolumeSpecName "kube-api-access-x9ksn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.329644 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-scripts" (OuterVolumeSpecName: "scripts") pod "57e9191a-aa60-498f-b621-2b83a7b795ed" (UID: "57e9191a-aa60-498f-b621-2b83a7b795ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.360239 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-config-data" (OuterVolumeSpecName: "config-data") pod "57e9191a-aa60-498f-b621-2b83a7b795ed" (UID: "57e9191a-aa60-498f-b621-2b83a7b795ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.379108 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57e9191a-aa60-498f-b621-2b83a7b795ed" (UID: "57e9191a-aa60-498f-b621-2b83a7b795ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.425128 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.425155 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9ksn\" (UniqueName: \"kubernetes.io/projected/57e9191a-aa60-498f-b621-2b83a7b795ed-kube-api-access-x9ksn\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.425165 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.425174 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/57e9191a-aa60-498f-b621-2b83a7b795ed-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.526449 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-nb\") pod \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\" (UID: \"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf\") " Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.527183 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" (UID: "a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.527672 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.690084 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-wdjvg" event={"ID":"57e9191a-aa60-498f-b621-2b83a7b795ed","Type":"ContainerDied","Data":"b0dafdbcd05230a453dd70f523b2192d0095ff1eb306414336b457a11e4822a0"} Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.690157 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0dafdbcd05230a453dd70f523b2192d0095ff1eb306414336b457a11e4822a0" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.690089 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-wdjvg" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.691943 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-tpdlf" event={"ID":"03d77cd8-31b2-4500-92c3-55ca3c0d09de","Type":"ContainerDied","Data":"e1e1020725fa1e5fb6d972b198c85497b37071ea30551b899c106fb157defdb4"} Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.691993 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1e1020725fa1e5fb6d972b198c85497b37071ea30551b899c106fb157defdb4" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.692132 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-tpdlf" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.695254 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" event={"ID":"a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf","Type":"ContainerDied","Data":"87980b017fba990245e4a878da2ccd87fdc784a9b137d63f5e404593c6179313"} Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.695284 5025 scope.go:117] "RemoveContainer" containerID="3b78cd90ae0947943209e117223a1b25701dc2b8bd997f4eee6275518e3db983" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.696443 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-89zdw" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.696479 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.739306 5025 scope.go:117] "RemoveContainer" containerID="0c7c4e3c6c52fdb30cbf761a3525873752d10706c6a678374babc0582477ad78" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.770477 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-89zdw"] Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.823345 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-89zdw"] Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.823388 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 10:54:22 crc kubenswrapper[5025]: E1004 10:54:22.844132 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" containerName="init" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.844170 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" containerName="init" Oct 04 10:54:22 crc kubenswrapper[5025]: E1004 10:54:22.844189 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e9191a-aa60-498f-b621-2b83a7b795ed" containerName="nova-cell1-conductor-db-sync" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.844195 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e9191a-aa60-498f-b621-2b83a7b795ed" containerName="nova-cell1-conductor-db-sync" Oct 04 10:54:22 crc kubenswrapper[5025]: E1004 10:54:22.844210 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d77cd8-31b2-4500-92c3-55ca3c0d09de" containerName="nova-manage" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.844216 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d77cd8-31b2-4500-92c3-55ca3c0d09de" containerName="nova-manage" Oct 04 10:54:22 crc kubenswrapper[5025]: E1004 10:54:22.844224 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" containerName="dnsmasq-dns" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.844229 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" containerName="dnsmasq-dns" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.844412 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="03d77cd8-31b2-4500-92c3-55ca3c0d09de" containerName="nova-manage" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.844429 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" containerName="dnsmasq-dns" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.844443 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e9191a-aa60-498f-b621-2b83a7b795ed" containerName="nova-cell1-conductor-db-sync" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.845044 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.846661 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.868318 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.896172 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.896402 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerName="nova-api-log" containerID="cri-o://bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f" gracePeriod=30 Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.896793 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerName="nova-api-api" containerID="cri-o://a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d" gracePeriod=30 Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.899213 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.949034 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc5d59e-f1c0-4b4a-ac69-326ffff2038a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.949091 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc5d59e-f1c0-4b4a-ac69-326ffff2038a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:22 crc kubenswrapper[5025]: I1004 10:54:22.949114 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgkwf\" (UniqueName: \"kubernetes.io/projected/ddc5d59e-f1c0-4b4a-ac69-326ffff2038a-kube-api-access-qgkwf\") pod \"nova-cell1-conductor-0\" (UID: \"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.050186 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc5d59e-f1c0-4b4a-ac69-326ffff2038a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.050239 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc5d59e-f1c0-4b4a-ac69-326ffff2038a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.050262 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgkwf\" (UniqueName: \"kubernetes.io/projected/ddc5d59e-f1c0-4b4a-ac69-326ffff2038a-kube-api-access-qgkwf\") pod \"nova-cell1-conductor-0\" (UID: \"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.058065 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc5d59e-f1c0-4b4a-ac69-326ffff2038a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.062633 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ddc5d59e-f1c0-4b4a-ac69-326ffff2038a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.072434 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgkwf\" (UniqueName: \"kubernetes.io/projected/ddc5d59e-f1c0-4b4a-ac69-326ffff2038a-kube-api-access-qgkwf\") pod \"nova-cell1-conductor-0\" (UID: \"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a\") " pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.208258 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.332132 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.705592 5025 generic.go:334] "Generic (PLEG): container finished" podID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerID="bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f" exitCode=143 Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.705684 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e6a4f0f-8574-434d-93d0-79a40c76776a","Type":"ContainerDied","Data":"bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f"} Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.706347 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="37905f14-a616-4f33-854d-e7313dd23b5d" containerName="nova-scheduler-scheduler" containerID="cri-o://a245353f4935b5ca0f2970b2560b03ee4e2a18645bd6a59025e321ee7263dc7b" gracePeriod=30 Oct 04 10:54:23 crc kubenswrapper[5025]: I1004 10:54:23.749512 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 04 10:54:23 crc kubenswrapper[5025]: W1004 10:54:23.753621 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddc5d59e_f1c0_4b4a_ac69_326ffff2038a.slice/crio-9054814bf05319e99f90c6341b7a71ce1e7e691b63b03647d1e4e5f3654eb684 WatchSource:0}: Error finding container 9054814bf05319e99f90c6341b7a71ce1e7e691b63b03647d1e4e5f3654eb684: Status 404 returned error can't find the container with id 9054814bf05319e99f90c6341b7a71ce1e7e691b63b03647d1e4e5f3654eb684 Oct 04 10:54:24 crc kubenswrapper[5025]: I1004 10:54:24.423873 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf" path="/var/lib/kubelet/pods/a8aa8ec5-3a31-4f9b-8b0c-8c45cd285caf/volumes" Oct 04 10:54:24 crc kubenswrapper[5025]: I1004 10:54:24.715686 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a","Type":"ContainerStarted","Data":"8e474c327a73648cedaf6307c202c4dbc51ab3179864bb65e827683e8d004f9a"} Oct 04 10:54:24 crc kubenswrapper[5025]: I1004 10:54:24.715725 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ddc5d59e-f1c0-4b4a-ac69-326ffff2038a","Type":"ContainerStarted","Data":"9054814bf05319e99f90c6341b7a71ce1e7e691b63b03647d1e4e5f3654eb684"} Oct 04 10:54:24 crc kubenswrapper[5025]: I1004 10:54:24.715814 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:24 crc kubenswrapper[5025]: I1004 10:54:24.744479 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.744457235 podStartE2EDuration="2.744457235s" podCreationTimestamp="2025-10-04 10:54:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:24.732158905 +0000 UTC m=+1193.157125805" watchObservedRunningTime="2025-10-04 10:54:24.744457235 +0000 UTC m=+1193.169424125" Oct 04 10:54:25 crc kubenswrapper[5025]: E1004 10:54:25.672871 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a245353f4935b5ca0f2970b2560b03ee4e2a18645bd6a59025e321ee7263dc7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 10:54:25 crc kubenswrapper[5025]: E1004 10:54:25.674915 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a245353f4935b5ca0f2970b2560b03ee4e2a18645bd6a59025e321ee7263dc7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 10:54:25 crc kubenswrapper[5025]: E1004 10:54:25.679674 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a245353f4935b5ca0f2970b2560b03ee4e2a18645bd6a59025e321ee7263dc7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 10:54:25 crc kubenswrapper[5025]: E1004 10:54:25.679721 5025 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="37905f14-a616-4f33-854d-e7313dd23b5d" containerName="nova-scheduler-scheduler" Oct 04 10:54:27 crc kubenswrapper[5025]: I1004 10:54:27.773338 5025 generic.go:334] "Generic (PLEG): container finished" podID="37905f14-a616-4f33-854d-e7313dd23b5d" containerID="a245353f4935b5ca0f2970b2560b03ee4e2a18645bd6a59025e321ee7263dc7b" exitCode=0 Oct 04 10:54:27 crc kubenswrapper[5025]: I1004 10:54:27.773434 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37905f14-a616-4f33-854d-e7313dd23b5d","Type":"ContainerDied","Data":"a245353f4935b5ca0f2970b2560b03ee4e2a18645bd6a59025e321ee7263dc7b"} Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.142363 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.242025 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs4df\" (UniqueName: \"kubernetes.io/projected/37905f14-a616-4f33-854d-e7313dd23b5d-kube-api-access-rs4df\") pod \"37905f14-a616-4f33-854d-e7313dd23b5d\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.242146 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-config-data\") pod \"37905f14-a616-4f33-854d-e7313dd23b5d\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.242184 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-combined-ca-bundle\") pod \"37905f14-a616-4f33-854d-e7313dd23b5d\" (UID: \"37905f14-a616-4f33-854d-e7313dd23b5d\") " Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.253256 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37905f14-a616-4f33-854d-e7313dd23b5d-kube-api-access-rs4df" (OuterVolumeSpecName: "kube-api-access-rs4df") pod "37905f14-a616-4f33-854d-e7313dd23b5d" (UID: "37905f14-a616-4f33-854d-e7313dd23b5d"). InnerVolumeSpecName "kube-api-access-rs4df". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.262922 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.273117 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-config-data" (OuterVolumeSpecName: "config-data") pod "37905f14-a616-4f33-854d-e7313dd23b5d" (UID: "37905f14-a616-4f33-854d-e7313dd23b5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.299984 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37905f14-a616-4f33-854d-e7313dd23b5d" (UID: "37905f14-a616-4f33-854d-e7313dd23b5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.344565 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs4df\" (UniqueName: \"kubernetes.io/projected/37905f14-a616-4f33-854d-e7313dd23b5d-kube-api-access-rs4df\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.344900 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.345043 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37905f14-a616-4f33-854d-e7313dd23b5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.715840 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.782767 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.782778 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37905f14-a616-4f33-854d-e7313dd23b5d","Type":"ContainerDied","Data":"18c3c54a566058a9b103dc5f385857554677d30b69777c2fbcdaf1cfa50eb211"} Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.782829 5025 scope.go:117] "RemoveContainer" containerID="a245353f4935b5ca0f2970b2560b03ee4e2a18645bd6a59025e321ee7263dc7b" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.784776 5025 generic.go:334] "Generic (PLEG): container finished" podID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerID="a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d" exitCode=0 Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.784801 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e6a4f0f-8574-434d-93d0-79a40c76776a","Type":"ContainerDied","Data":"a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d"} Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.784822 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2e6a4f0f-8574-434d-93d0-79a40c76776a","Type":"ContainerDied","Data":"fc0c66196791fb43ed25ef9a13321f53c3f55f744d71c09394f1f7b3b6ebbf49"} Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.784897 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.810811 5025 scope.go:117] "RemoveContainer" containerID="a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.829560 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.838058 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.840161 5025 scope.go:117] "RemoveContainer" containerID="bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.846599 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:54:28 crc kubenswrapper[5025]: E1004 10:54:28.847201 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37905f14-a616-4f33-854d-e7313dd23b5d" containerName="nova-scheduler-scheduler" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.847225 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="37905f14-a616-4f33-854d-e7313dd23b5d" containerName="nova-scheduler-scheduler" Oct 04 10:54:28 crc kubenswrapper[5025]: E1004 10:54:28.847277 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerName="nova-api-log" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.847286 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerName="nova-api-log" Oct 04 10:54:28 crc kubenswrapper[5025]: E1004 10:54:28.847298 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerName="nova-api-api" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.847305 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerName="nova-api-api" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.847520 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerName="nova-api-log" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.847534 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" containerName="nova-api-api" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.847545 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="37905f14-a616-4f33-854d-e7313dd23b5d" containerName="nova-scheduler-scheduler" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.848337 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.850182 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.854491 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-config-data\") pod \"2e6a4f0f-8574-434d-93d0-79a40c76776a\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.854577 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bzhw\" (UniqueName: \"kubernetes.io/projected/2e6a4f0f-8574-434d-93d0-79a40c76776a-kube-api-access-9bzhw\") pod \"2e6a4f0f-8574-434d-93d0-79a40c76776a\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.854657 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-combined-ca-bundle\") pod \"2e6a4f0f-8574-434d-93d0-79a40c76776a\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.854724 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e6a4f0f-8574-434d-93d0-79a40c76776a-logs\") pod \"2e6a4f0f-8574-434d-93d0-79a40c76776a\" (UID: \"2e6a4f0f-8574-434d-93d0-79a40c76776a\") " Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.855663 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e6a4f0f-8574-434d-93d0-79a40c76776a-logs" (OuterVolumeSpecName: "logs") pod "2e6a4f0f-8574-434d-93d0-79a40c76776a" (UID: "2e6a4f0f-8574-434d-93d0-79a40c76776a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.856220 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.859716 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e6a4f0f-8574-434d-93d0-79a40c76776a-kube-api-access-9bzhw" (OuterVolumeSpecName: "kube-api-access-9bzhw") pod "2e6a4f0f-8574-434d-93d0-79a40c76776a" (UID: "2e6a4f0f-8574-434d-93d0-79a40c76776a"). InnerVolumeSpecName "kube-api-access-9bzhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.864689 5025 scope.go:117] "RemoveContainer" containerID="a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d" Oct 04 10:54:28 crc kubenswrapper[5025]: E1004 10:54:28.865263 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d\": container with ID starting with a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d not found: ID does not exist" containerID="a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.865304 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d"} err="failed to get container status \"a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d\": rpc error: code = NotFound desc = could not find container \"a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d\": container with ID starting with a2cf279c1800d6a02fb58fd81b87c2dc6c6446330199cc8b552a1a20c735231d not found: ID does not exist" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.865331 5025 scope.go:117] "RemoveContainer" containerID="bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f" Oct 04 10:54:28 crc kubenswrapper[5025]: E1004 10:54:28.865663 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f\": container with ID starting with bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f not found: ID does not exist" containerID="bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.865688 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f"} err="failed to get container status \"bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f\": rpc error: code = NotFound desc = could not find container \"bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f\": container with ID starting with bed8667fd8df9ae24f87701bc65d620d7f8bb704cf753fe7763e0d538849947f not found: ID does not exist" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.883258 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-config-data" (OuterVolumeSpecName: "config-data") pod "2e6a4f0f-8574-434d-93d0-79a40c76776a" (UID: "2e6a4f0f-8574-434d-93d0-79a40c76776a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.884840 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e6a4f0f-8574-434d-93d0-79a40c76776a" (UID: "2e6a4f0f-8574-434d-93d0-79a40c76776a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.956719 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-config-data\") pod \"nova-scheduler-0\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.956795 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdfrb\" (UniqueName: \"kubernetes.io/projected/e057802b-4b1c-493a-8ca1-15f81378a8fc-kube-api-access-kdfrb\") pod \"nova-scheduler-0\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.957106 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.957452 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.957481 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bzhw\" (UniqueName: \"kubernetes.io/projected/2e6a4f0f-8574-434d-93d0-79a40c76776a-kube-api-access-9bzhw\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.957498 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e6a4f0f-8574-434d-93d0-79a40c76776a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:28 crc kubenswrapper[5025]: I1004 10:54:28.957511 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e6a4f0f-8574-434d-93d0-79a40c76776a-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.059447 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-config-data\") pod \"nova-scheduler-0\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.059536 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdfrb\" (UniqueName: \"kubernetes.io/projected/e057802b-4b1c-493a-8ca1-15f81378a8fc-kube-api-access-kdfrb\") pod \"nova-scheduler-0\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.059640 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.064174 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.064610 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-config-data\") pod \"nova-scheduler-0\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.080384 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdfrb\" (UniqueName: \"kubernetes.io/projected/e057802b-4b1c-493a-8ca1-15f81378a8fc-kube-api-access-kdfrb\") pod \"nova-scheduler-0\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " pod="openstack/nova-scheduler-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.168836 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.176483 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.193664 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.195483 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.204499 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.205521 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.246275 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.263746 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.264048 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-config-data\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.264178 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43ca7288-b2d5-4df8-9776-677efaf39087-logs\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.264328 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kfhm\" (UniqueName: \"kubernetes.io/projected/43ca7288-b2d5-4df8-9776-677efaf39087-kube-api-access-5kfhm\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.366438 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.366609 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-config-data\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.366639 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43ca7288-b2d5-4df8-9776-677efaf39087-logs\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.366716 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kfhm\" (UniqueName: \"kubernetes.io/projected/43ca7288-b2d5-4df8-9776-677efaf39087-kube-api-access-5kfhm\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.367380 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43ca7288-b2d5-4df8-9776-677efaf39087-logs\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.373324 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-config-data\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.373834 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.388517 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kfhm\" (UniqueName: \"kubernetes.io/projected/43ca7288-b2d5-4df8-9776-677efaf39087-kube-api-access-5kfhm\") pod \"nova-api-0\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.521669 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.799910 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:54:29 crc kubenswrapper[5025]: W1004 10:54:29.811656 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode057802b_4b1c_493a_8ca1_15f81378a8fc.slice/crio-092a662ad4ec84e1d3d30131a517ef463ab15968bed94709a4076794f8c2fff9 WatchSource:0}: Error finding container 092a662ad4ec84e1d3d30131a517ef463ab15968bed94709a4076794f8c2fff9: Status 404 returned error can't find the container with id 092a662ad4ec84e1d3d30131a517ef463ab15968bed94709a4076794f8c2fff9 Oct 04 10:54:29 crc kubenswrapper[5025]: I1004 10:54:29.964582 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:29 crc kubenswrapper[5025]: W1004 10:54:29.972534 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43ca7288_b2d5_4df8_9776_677efaf39087.slice/crio-f9f0aee9c147af6b2a00c295421c57b2667478657bbddaa5168a83526ae06927 WatchSource:0}: Error finding container f9f0aee9c147af6b2a00c295421c57b2667478657bbddaa5168a83526ae06927: Status 404 returned error can't find the container with id f9f0aee9c147af6b2a00c295421c57b2667478657bbddaa5168a83526ae06927 Oct 04 10:54:30 crc kubenswrapper[5025]: I1004 10:54:30.428450 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e6a4f0f-8574-434d-93d0-79a40c76776a" path="/var/lib/kubelet/pods/2e6a4f0f-8574-434d-93d0-79a40c76776a/volumes" Oct 04 10:54:30 crc kubenswrapper[5025]: I1004 10:54:30.429975 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37905f14-a616-4f33-854d-e7313dd23b5d" path="/var/lib/kubelet/pods/37905f14-a616-4f33-854d-e7313dd23b5d/volumes" Oct 04 10:54:30 crc kubenswrapper[5025]: I1004 10:54:30.816250 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"43ca7288-b2d5-4df8-9776-677efaf39087","Type":"ContainerStarted","Data":"665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5"} Oct 04 10:54:30 crc kubenswrapper[5025]: I1004 10:54:30.816291 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"43ca7288-b2d5-4df8-9776-677efaf39087","Type":"ContainerStarted","Data":"54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651"} Oct 04 10:54:30 crc kubenswrapper[5025]: I1004 10:54:30.816302 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"43ca7288-b2d5-4df8-9776-677efaf39087","Type":"ContainerStarted","Data":"f9f0aee9c147af6b2a00c295421c57b2667478657bbddaa5168a83526ae06927"} Oct 04 10:54:30 crc kubenswrapper[5025]: I1004 10:54:30.817676 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e057802b-4b1c-493a-8ca1-15f81378a8fc","Type":"ContainerStarted","Data":"401857d48166fab6f580f3e80942cc354d7183f425fbe92d2b14f9b1a07390da"} Oct 04 10:54:30 crc kubenswrapper[5025]: I1004 10:54:30.817736 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e057802b-4b1c-493a-8ca1-15f81378a8fc","Type":"ContainerStarted","Data":"092a662ad4ec84e1d3d30131a517ef463ab15968bed94709a4076794f8c2fff9"} Oct 04 10:54:30 crc kubenswrapper[5025]: I1004 10:54:30.842151 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.842129377 podStartE2EDuration="1.842129377s" podCreationTimestamp="2025-10-04 10:54:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:30.836573034 +0000 UTC m=+1199.261539904" watchObservedRunningTime="2025-10-04 10:54:30.842129377 +0000 UTC m=+1199.267096257" Oct 04 10:54:30 crc kubenswrapper[5025]: I1004 10:54:30.864279 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.864259749 podStartE2EDuration="2.864259749s" podCreationTimestamp="2025-10-04 10:54:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:30.858792428 +0000 UTC m=+1199.283759308" watchObservedRunningTime="2025-10-04 10:54:30.864259749 +0000 UTC m=+1199.289226629" Oct 04 10:54:34 crc kubenswrapper[5025]: I1004 10:54:34.247661 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 10:54:39 crc kubenswrapper[5025]: I1004 10:54:39.247608 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 10:54:39 crc kubenswrapper[5025]: I1004 10:54:39.286224 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 10:54:39 crc kubenswrapper[5025]: I1004 10:54:39.523693 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 10:54:39 crc kubenswrapper[5025]: I1004 10:54:39.523763 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 10:54:39 crc kubenswrapper[5025]: I1004 10:54:39.944381 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 10:54:40 crc kubenswrapper[5025]: I1004 10:54:40.609229 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 10:54:40 crc kubenswrapper[5025]: I1004 10:54:40.609276 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 04 10:54:44 crc kubenswrapper[5025]: I1004 10:54:44.713126 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:54:44 crc kubenswrapper[5025]: I1004 10:54:44.713524 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.008842 5025 generic.go:334] "Generic (PLEG): container finished" podID="dc28e385-af83-45f7-9bb3-25edf85e2866" containerID="14743e2481323cb7ef3ffe0099e6662bca90c3c14cca7c7625b0dbd06bbaa8f6" exitCode=137 Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.009532 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dc28e385-af83-45f7-9bb3-25edf85e2866","Type":"ContainerDied","Data":"14743e2481323cb7ef3ffe0099e6662bca90c3c14cca7c7625b0dbd06bbaa8f6"} Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.009568 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"dc28e385-af83-45f7-9bb3-25edf85e2866","Type":"ContainerDied","Data":"06d959aef2a731c121680f729a7e5f1c75879fe86ac5b9de722554fd8805cd53"} Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.009587 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06d959aef2a731c121680f729a7e5f1c75879fe86ac5b9de722554fd8805cd53" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.012972 5025 generic.go:334] "Generic (PLEG): container finished" podID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" containerID="08c5ac95c8a494b2c84c8f29df3f940bd79ee23fee759ecba9b6c7c8ac4066d7" exitCode=137 Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.013003 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec9f4766-d4bf-4db8-851e-368d670a5fa7","Type":"ContainerDied","Data":"08c5ac95c8a494b2c84c8f29df3f940bd79ee23fee759ecba9b6c7c8ac4066d7"} Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.013075 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ec9f4766-d4bf-4db8-851e-368d670a5fa7","Type":"ContainerDied","Data":"a4fae44b981a747f25944aec6124810595ab50d0b3bd62e5b1acdc7c1bffe7e9"} Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.013090 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4fae44b981a747f25944aec6124810595ab50d0b3bd62e5b1acdc7c1bffe7e9" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.013699 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.026827 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.045823 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.125634 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-config-data\") pod \"dc28e385-af83-45f7-9bb3-25edf85e2866\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.126709 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-config-data\") pod \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.126866 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-combined-ca-bundle\") pod \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.126936 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5btbk\" (UniqueName: \"kubernetes.io/projected/ec9f4766-d4bf-4db8-851e-368d670a5fa7-kube-api-access-5btbk\") pod \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.126999 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-combined-ca-bundle\") pod \"dc28e385-af83-45f7-9bb3-25edf85e2866\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.127054 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6qvq\" (UniqueName: \"kubernetes.io/projected/dc28e385-af83-45f7-9bb3-25edf85e2866-kube-api-access-w6qvq\") pod \"dc28e385-af83-45f7-9bb3-25edf85e2866\" (UID: \"dc28e385-af83-45f7-9bb3-25edf85e2866\") " Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.127120 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec9f4766-d4bf-4db8-851e-368d670a5fa7-logs\") pod \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\" (UID: \"ec9f4766-d4bf-4db8-851e-368d670a5fa7\") " Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.129482 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec9f4766-d4bf-4db8-851e-368d670a5fa7-logs" (OuterVolumeSpecName: "logs") pod "ec9f4766-d4bf-4db8-851e-368d670a5fa7" (UID: "ec9f4766-d4bf-4db8-851e-368d670a5fa7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.133207 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec9f4766-d4bf-4db8-851e-368d670a5fa7-kube-api-access-5btbk" (OuterVolumeSpecName: "kube-api-access-5btbk") pod "ec9f4766-d4bf-4db8-851e-368d670a5fa7" (UID: "ec9f4766-d4bf-4db8-851e-368d670a5fa7"). InnerVolumeSpecName "kube-api-access-5btbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.134253 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc28e385-af83-45f7-9bb3-25edf85e2866-kube-api-access-w6qvq" (OuterVolumeSpecName: "kube-api-access-w6qvq") pod "dc28e385-af83-45f7-9bb3-25edf85e2866" (UID: "dc28e385-af83-45f7-9bb3-25edf85e2866"). InnerVolumeSpecName "kube-api-access-w6qvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.166275 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc28e385-af83-45f7-9bb3-25edf85e2866" (UID: "dc28e385-af83-45f7-9bb3-25edf85e2866"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.169150 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec9f4766-d4bf-4db8-851e-368d670a5fa7" (UID: "ec9f4766-d4bf-4db8-851e-368d670a5fa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.199181 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-config-data" (OuterVolumeSpecName: "config-data") pod "ec9f4766-d4bf-4db8-851e-368d670a5fa7" (UID: "ec9f4766-d4bf-4db8-851e-368d670a5fa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.209206 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-config-data" (OuterVolumeSpecName: "config-data") pod "dc28e385-af83-45f7-9bb3-25edf85e2866" (UID: "dc28e385-af83-45f7-9bb3-25edf85e2866"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.230233 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.230289 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5btbk\" (UniqueName: \"kubernetes.io/projected/ec9f4766-d4bf-4db8-851e-368d670a5fa7-kube-api-access-5btbk\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.230308 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.230320 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6qvq\" (UniqueName: \"kubernetes.io/projected/dc28e385-af83-45f7-9bb3-25edf85e2866-kube-api-access-w6qvq\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.230330 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ec9f4766-d4bf-4db8-851e-368d670a5fa7-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.230341 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc28e385-af83-45f7-9bb3-25edf85e2866-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:47 crc kubenswrapper[5025]: I1004 10:54:47.230350 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec9f4766-d4bf-4db8-851e-368d670a5fa7-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.023775 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.023796 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.060791 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.083680 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.096137 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.113419 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.129454 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:54:48 crc kubenswrapper[5025]: E1004 10:54:48.129909 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" containerName="nova-metadata-log" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.129933 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" containerName="nova-metadata-log" Oct 04 10:54:48 crc kubenswrapper[5025]: E1004 10:54:48.129960 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc28e385-af83-45f7-9bb3-25edf85e2866" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.129968 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc28e385-af83-45f7-9bb3-25edf85e2866" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 10:54:48 crc kubenswrapper[5025]: E1004 10:54:48.129986 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" containerName="nova-metadata-metadata" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.129992 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" containerName="nova-metadata-metadata" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.130170 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc28e385-af83-45f7-9bb3-25edf85e2866" containerName="nova-cell1-novncproxy-novncproxy" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.130182 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" containerName="nova-metadata-log" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.130196 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" containerName="nova-metadata-metadata" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.131553 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.134156 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.138536 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.141840 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.151926 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.153318 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.155112 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.155392 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.156134 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.165504 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.250701 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.250764 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckc2m\" (UniqueName: \"kubernetes.io/projected/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-kube-api-access-ckc2m\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.250808 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.250878 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a195e5b2-84f4-4f36-9f73-0c0b30262c06-logs\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.250940 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkx6k\" (UniqueName: \"kubernetes.io/projected/a195e5b2-84f4-4f36-9f73-0c0b30262c06-kube-api-access-fkx6k\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.250971 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.250993 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.251096 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-config-data\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.251167 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.251359 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.353388 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckc2m\" (UniqueName: \"kubernetes.io/projected/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-kube-api-access-ckc2m\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.353454 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.353482 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a195e5b2-84f4-4f36-9f73-0c0b30262c06-logs\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.353556 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkx6k\" (UniqueName: \"kubernetes.io/projected/a195e5b2-84f4-4f36-9f73-0c0b30262c06-kube-api-access-fkx6k\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.353592 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.353622 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.353644 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-config-data\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.353682 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.353737 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.353800 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.356472 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a195e5b2-84f4-4f36-9f73-0c0b30262c06-logs\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.358790 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.359131 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.360146 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.360369 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.361653 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-config-data\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.361701 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.363147 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.372909 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckc2m\" (UniqueName: \"kubernetes.io/projected/5e1bccb3-569f-45d0-a778-0fd8b04f8cca-kube-api-access-ckc2m\") pod \"nova-cell1-novncproxy-0\" (UID: \"5e1bccb3-569f-45d0-a778-0fd8b04f8cca\") " pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.383560 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkx6k\" (UniqueName: \"kubernetes.io/projected/a195e5b2-84f4-4f36-9f73-0c0b30262c06-kube-api-access-fkx6k\") pod \"nova-metadata-0\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.425889 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc28e385-af83-45f7-9bb3-25edf85e2866" path="/var/lib/kubelet/pods/dc28e385-af83-45f7-9bb3-25edf85e2866/volumes" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.427246 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec9f4766-d4bf-4db8-851e-368d670a5fa7" path="/var/lib/kubelet/pods/ec9f4766-d4bf-4db8-851e-368d670a5fa7/volumes" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.451433 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 10:54:48 crc kubenswrapper[5025]: I1004 10:54:48.473628 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:49 crc kubenswrapper[5025]: W1004 10:54:49.003241 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda195e5b2_84f4_4f36_9f73_0c0b30262c06.slice/crio-2f1e4cc45e5d2276b489a8b68727071b43c9f9e82d4370579ae06c2ef53ec282 WatchSource:0}: Error finding container 2f1e4cc45e5d2276b489a8b68727071b43c9f9e82d4370579ae06c2ef53ec282: Status 404 returned error can't find the container with id 2f1e4cc45e5d2276b489a8b68727071b43c9f9e82d4370579ae06c2ef53ec282 Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.005164 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.033958 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a195e5b2-84f4-4f36-9f73-0c0b30262c06","Type":"ContainerStarted","Data":"2f1e4cc45e5d2276b489a8b68727071b43c9f9e82d4370579ae06c2ef53ec282"} Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.064467 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 04 10:54:49 crc kubenswrapper[5025]: W1004 10:54:49.066664 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e1bccb3_569f_45d0_a778_0fd8b04f8cca.slice/crio-282cfe4c5edd702751e4ff0cd8b3fff2a35a157e1c55bdb7e6c07d8794287cc5 WatchSource:0}: Error finding container 282cfe4c5edd702751e4ff0cd8b3fff2a35a157e1c55bdb7e6c07d8794287cc5: Status 404 returned error can't find the container with id 282cfe4c5edd702751e4ff0cd8b3fff2a35a157e1c55bdb7e6c07d8794287cc5 Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.528125 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.528503 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.529186 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.529224 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.534291 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.544458 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.751555 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-glssw"] Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.753224 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.771204 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-glssw"] Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.892897 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.892972 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.893006 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.893082 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7zgx\" (UniqueName: \"kubernetes.io/projected/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-kube-api-access-f7zgx\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.893115 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-config\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.893170 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.995294 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.995441 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.995484 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.995515 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.995546 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7zgx\" (UniqueName: \"kubernetes.io/projected/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-kube-api-access-f7zgx\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.995575 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-config\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.996306 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.996395 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.996395 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.996550 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-config\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:49 crc kubenswrapper[5025]: I1004 10:54:49.996841 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:50 crc kubenswrapper[5025]: I1004 10:54:50.028568 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7zgx\" (UniqueName: \"kubernetes.io/projected/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-kube-api-access-f7zgx\") pod \"dnsmasq-dns-5c7b6c5df9-glssw\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:50 crc kubenswrapper[5025]: I1004 10:54:50.049847 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e1bccb3-569f-45d0-a778-0fd8b04f8cca","Type":"ContainerStarted","Data":"4c7de1fe66954011b7f3ebb93b9205fdf73520b166c6a35a6a4770c562819bc2"} Oct 04 10:54:50 crc kubenswrapper[5025]: I1004 10:54:50.051385 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5e1bccb3-569f-45d0-a778-0fd8b04f8cca","Type":"ContainerStarted","Data":"282cfe4c5edd702751e4ff0cd8b3fff2a35a157e1c55bdb7e6c07d8794287cc5"} Oct 04 10:54:50 crc kubenswrapper[5025]: I1004 10:54:50.052747 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a195e5b2-84f4-4f36-9f73-0c0b30262c06","Type":"ContainerStarted","Data":"f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694"} Oct 04 10:54:50 crc kubenswrapper[5025]: I1004 10:54:50.052786 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a195e5b2-84f4-4f36-9f73-0c0b30262c06","Type":"ContainerStarted","Data":"b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1"} Oct 04 10:54:50 crc kubenswrapper[5025]: I1004 10:54:50.073721 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:50 crc kubenswrapper[5025]: I1004 10:54:50.081745 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.081723104 podStartE2EDuration="2.081723104s" podCreationTimestamp="2025-10-04 10:54:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:50.077281831 +0000 UTC m=+1218.502248711" watchObservedRunningTime="2025-10-04 10:54:50.081723104 +0000 UTC m=+1218.506689984" Oct 04 10:54:50 crc kubenswrapper[5025]: I1004 10:54:50.117030 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.116986608 podStartE2EDuration="2.116986608s" podCreationTimestamp="2025-10-04 10:54:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:50.105193332 +0000 UTC m=+1218.530160212" watchObservedRunningTime="2025-10-04 10:54:50.116986608 +0000 UTC m=+1218.541953488" Oct 04 10:54:50 crc kubenswrapper[5025]: I1004 10:54:50.660638 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-glssw"] Oct 04 10:54:51 crc kubenswrapper[5025]: I1004 10:54:51.062309 5025 generic.go:334] "Generic (PLEG): container finished" podID="1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" containerID="fbf2ccaacec2c80fd6c1c7a5359b0484038b8df2161f7c15cf5192431f49bb6a" exitCode=0 Oct 04 10:54:51 crc kubenswrapper[5025]: I1004 10:54:51.062355 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" event={"ID":"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f","Type":"ContainerDied","Data":"fbf2ccaacec2c80fd6c1c7a5359b0484038b8df2161f7c15cf5192431f49bb6a"} Oct 04 10:54:51 crc kubenswrapper[5025]: I1004 10:54:51.062697 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" event={"ID":"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f","Type":"ContainerStarted","Data":"9d8e15c12f0b77325c1ff594615d3a87ea5b040e47ef6967dab636026ded6574"} Oct 04 10:54:51 crc kubenswrapper[5025]: I1004 10:54:51.781676 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:51 crc kubenswrapper[5025]: I1004 10:54:51.782231 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="ceilometer-central-agent" containerID="cri-o://6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076" gracePeriod=30 Oct 04 10:54:51 crc kubenswrapper[5025]: I1004 10:54:51.782364 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="proxy-httpd" containerID="cri-o://4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0" gracePeriod=30 Oct 04 10:54:51 crc kubenswrapper[5025]: I1004 10:54:51.782373 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="ceilometer-notification-agent" containerID="cri-o://0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120" gracePeriod=30 Oct 04 10:54:51 crc kubenswrapper[5025]: I1004 10:54:51.782335 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="sg-core" containerID="cri-o://714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073" gracePeriod=30 Oct 04 10:54:52 crc kubenswrapper[5025]: I1004 10:54:52.075057 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" event={"ID":"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f","Type":"ContainerStarted","Data":"c46b538b3cb2c39640c9a6eb939dd6d80c04676def70e3fd0e24c4673f2c6bbb"} Oct 04 10:54:52 crc kubenswrapper[5025]: I1004 10:54:52.076195 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:54:52 crc kubenswrapper[5025]: I1004 10:54:52.080188 5025 generic.go:334] "Generic (PLEG): container finished" podID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerID="4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0" exitCode=0 Oct 04 10:54:52 crc kubenswrapper[5025]: I1004 10:54:52.080213 5025 generic.go:334] "Generic (PLEG): container finished" podID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerID="714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073" exitCode=2 Oct 04 10:54:52 crc kubenswrapper[5025]: I1004 10:54:52.080705 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a","Type":"ContainerDied","Data":"4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0"} Oct 04 10:54:52 crc kubenswrapper[5025]: I1004 10:54:52.080730 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a","Type":"ContainerDied","Data":"714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073"} Oct 04 10:54:52 crc kubenswrapper[5025]: I1004 10:54:52.096850 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" podStartSLOduration=3.096838092 podStartE2EDuration="3.096838092s" podCreationTimestamp="2025-10-04 10:54:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:52.094454526 +0000 UTC m=+1220.519421406" watchObservedRunningTime="2025-10-04 10:54:52.096838092 +0000 UTC m=+1220.521804972" Oct 04 10:54:52 crc kubenswrapper[5025]: I1004 10:54:52.736868 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:52 crc kubenswrapper[5025]: I1004 10:54:52.738063 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" containerName="nova-api-api" containerID="cri-o://665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5" gracePeriod=30 Oct 04 10:54:52 crc kubenswrapper[5025]: I1004 10:54:52.737961 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" containerName="nova-api-log" containerID="cri-o://54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651" gracePeriod=30 Oct 04 10:54:53 crc kubenswrapper[5025]: I1004 10:54:53.089883 5025 generic.go:334] "Generic (PLEG): container finished" podID="43ca7288-b2d5-4df8-9776-677efaf39087" containerID="54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651" exitCode=143 Oct 04 10:54:53 crc kubenswrapper[5025]: I1004 10:54:53.089954 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"43ca7288-b2d5-4df8-9776-677efaf39087","Type":"ContainerDied","Data":"54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651"} Oct 04 10:54:53 crc kubenswrapper[5025]: I1004 10:54:53.092156 5025 generic.go:334] "Generic (PLEG): container finished" podID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerID="6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076" exitCode=0 Oct 04 10:54:53 crc kubenswrapper[5025]: I1004 10:54:53.092196 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a","Type":"ContainerDied","Data":"6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076"} Oct 04 10:54:53 crc kubenswrapper[5025]: I1004 10:54:53.452691 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 10:54:53 crc kubenswrapper[5025]: I1004 10:54:53.452750 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 10:54:53 crc kubenswrapper[5025]: I1004 10:54:53.474559 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.647599 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.720669 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43ca7288-b2d5-4df8-9776-677efaf39087-logs\") pod \"43ca7288-b2d5-4df8-9776-677efaf39087\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.720724 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kfhm\" (UniqueName: \"kubernetes.io/projected/43ca7288-b2d5-4df8-9776-677efaf39087-kube-api-access-5kfhm\") pod \"43ca7288-b2d5-4df8-9776-677efaf39087\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.720962 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-config-data\") pod \"43ca7288-b2d5-4df8-9776-677efaf39087\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.721124 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-combined-ca-bundle\") pod \"43ca7288-b2d5-4df8-9776-677efaf39087\" (UID: \"43ca7288-b2d5-4df8-9776-677efaf39087\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.721914 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43ca7288-b2d5-4df8-9776-677efaf39087-logs" (OuterVolumeSpecName: "logs") pod "43ca7288-b2d5-4df8-9776-677efaf39087" (UID: "43ca7288-b2d5-4df8-9776-677efaf39087"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.735210 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43ca7288-b2d5-4df8-9776-677efaf39087-kube-api-access-5kfhm" (OuterVolumeSpecName: "kube-api-access-5kfhm") pod "43ca7288-b2d5-4df8-9776-677efaf39087" (UID: "43ca7288-b2d5-4df8-9776-677efaf39087"). InnerVolumeSpecName "kube-api-access-5kfhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.747735 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "43ca7288-b2d5-4df8-9776-677efaf39087" (UID: "43ca7288-b2d5-4df8-9776-677efaf39087"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.755506 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-config-data" (OuterVolumeSpecName: "config-data") pod "43ca7288-b2d5-4df8-9776-677efaf39087" (UID: "43ca7288-b2d5-4df8-9776-677efaf39087"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.787745 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.823219 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-combined-ca-bundle\") pod \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.823295 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-run-httpd\") pod \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.823367 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-log-httpd\") pod \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.823429 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-config-data\") pod \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.823475 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-scripts\") pod \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.823519 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdcsz\" (UniqueName: \"kubernetes.io/projected/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-kube-api-access-cdcsz\") pod \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.823547 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-sg-core-conf-yaml\") pod \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.823593 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-ceilometer-tls-certs\") pod \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\" (UID: \"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a\") " Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.824100 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.824123 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43ca7288-b2d5-4df8-9776-677efaf39087-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.824139 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43ca7288-b2d5-4df8-9776-677efaf39087-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.824152 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kfhm\" (UniqueName: \"kubernetes.io/projected/43ca7288-b2d5-4df8-9776-677efaf39087-kube-api-access-5kfhm\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.825341 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" (UID: "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.825917 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" (UID: "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.827409 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-kube-api-access-cdcsz" (OuterVolumeSpecName: "kube-api-access-cdcsz") pod "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" (UID: "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a"). InnerVolumeSpecName "kube-api-access-cdcsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.828635 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-scripts" (OuterVolumeSpecName: "scripts") pod "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" (UID: "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.848584 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" (UID: "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.890167 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" (UID: "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.897926 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" (UID: "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.926009 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.926056 5025 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.926066 5025 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.926075 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.926084 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdcsz\" (UniqueName: \"kubernetes.io/projected/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-kube-api-access-cdcsz\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.926094 5025 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.926101 5025 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:56 crc kubenswrapper[5025]: I1004 10:54:56.955702 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-config-data" (OuterVolumeSpecName: "config-data") pod "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" (UID: "eee9436c-7c2b-4ccf-9218-1e2da6e96e6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.027900 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.125967 5025 generic.go:334] "Generic (PLEG): container finished" podID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerID="0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120" exitCode=0 Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.126058 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.126059 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a","Type":"ContainerDied","Data":"0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120"} Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.126163 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eee9436c-7c2b-4ccf-9218-1e2da6e96e6a","Type":"ContainerDied","Data":"99939ceedec07f8253f6e6d51e4a5cdf63094384a4259d9f7d230ed34f3fe4e8"} Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.126192 5025 scope.go:117] "RemoveContainer" containerID="4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.128146 5025 generic.go:334] "Generic (PLEG): container finished" podID="43ca7288-b2d5-4df8-9776-677efaf39087" containerID="665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5" exitCode=0 Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.128174 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"43ca7288-b2d5-4df8-9776-677efaf39087","Type":"ContainerDied","Data":"665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5"} Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.128202 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"43ca7288-b2d5-4df8-9776-677efaf39087","Type":"ContainerDied","Data":"f9f0aee9c147af6b2a00c295421c57b2667478657bbddaa5168a83526ae06927"} Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.128246 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.150973 5025 scope.go:117] "RemoveContainer" containerID="714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.171054 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.177820 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.184381 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.189124 5025 scope.go:117] "RemoveContainer" containerID="0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.197807 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.207125 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.207619 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="proxy-httpd" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.207644 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="proxy-httpd" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.207671 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="sg-core" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.207680 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="sg-core" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.207703 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" containerName="nova-api-log" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.207711 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" containerName="nova-api-log" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.207727 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="ceilometer-central-agent" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.207735 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="ceilometer-central-agent" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.207756 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="ceilometer-notification-agent" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.207764 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="ceilometer-notification-agent" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.207807 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" containerName="nova-api-api" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.207818 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" containerName="nova-api-api" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.208055 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="ceilometer-central-agent" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.208101 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="sg-core" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.208120 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" containerName="nova-api-log" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.208136 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="proxy-httpd" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.208152 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" containerName="ceilometer-notification-agent" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.208171 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" containerName="nova-api-api" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.210304 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.217305 5025 scope.go:117] "RemoveContainer" containerID="6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.217971 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.218265 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.218295 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.218368 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.219512 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.232710 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.233938 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.241952 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.247540 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-scripts\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.247679 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz6wg\" (UniqueName: \"kubernetes.io/projected/0baa8e1a-723d-49e2-9f65-0d95550a8633-kube-api-access-nz6wg\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.247716 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.247898 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.247963 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0baa8e1a-723d-49e2-9f65-0d95550a8633-log-httpd\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.247994 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-config-data\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.248108 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.248252 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0baa8e1a-723d-49e2-9f65-0d95550a8633-run-httpd\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.252434 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.254696 5025 scope.go:117] "RemoveContainer" containerID="4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.258647 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0\": container with ID starting with 4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0 not found: ID does not exist" containerID="4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.258703 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0"} err="failed to get container status \"4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0\": rpc error: code = NotFound desc = could not find container \"4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0\": container with ID starting with 4cfd910da0b741ab557636b1430dd82c53c6b105bbda93906dd36878ddb59ff0 not found: ID does not exist" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.258733 5025 scope.go:117] "RemoveContainer" containerID="714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.259192 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073\": container with ID starting with 714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073 not found: ID does not exist" containerID="714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.259238 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073"} err="failed to get container status \"714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073\": rpc error: code = NotFound desc = could not find container \"714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073\": container with ID starting with 714568490a56cd6779a035c63d33b8850304eabddc6d27c0837b8b5d67625073 not found: ID does not exist" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.259251 5025 scope.go:117] "RemoveContainer" containerID="0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.259668 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120\": container with ID starting with 0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120 not found: ID does not exist" containerID="0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.259725 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120"} err="failed to get container status \"0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120\": rpc error: code = NotFound desc = could not find container \"0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120\": container with ID starting with 0a839012be75214b0aa785d06ff6f5a9625a294c1148660ce7f9d01af2088120 not found: ID does not exist" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.259748 5025 scope.go:117] "RemoveContainer" containerID="6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.260102 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076\": container with ID starting with 6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076 not found: ID does not exist" containerID="6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.260123 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076"} err="failed to get container status \"6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076\": rpc error: code = NotFound desc = could not find container \"6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076\": container with ID starting with 6defeac8321ea4cabce28596117d2d3c529c5f17358720a4f1d9e952a473f076 not found: ID does not exist" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.260200 5025 scope.go:117] "RemoveContainer" containerID="665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.268167 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.336303 5025 scope.go:117] "RemoveContainer" containerID="54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350044 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz6wg\" (UniqueName: \"kubernetes.io/projected/0baa8e1a-723d-49e2-9f65-0d95550a8633-kube-api-access-nz6wg\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350091 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350147 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s52n7\" (UniqueName: \"kubernetes.io/projected/adec9469-e051-42e1-a46c-87ce7f54aab3-kube-api-access-s52n7\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350202 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-public-tls-certs\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350236 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350257 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350283 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0baa8e1a-723d-49e2-9f65-0d95550a8633-log-httpd\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350300 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-config-data\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350318 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350338 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adec9469-e051-42e1-a46c-87ce7f54aab3-logs\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350383 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0baa8e1a-723d-49e2-9f65-0d95550a8633-run-httpd\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350410 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-config-data\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350432 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.350471 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-scripts\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.351049 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0baa8e1a-723d-49e2-9f65-0d95550a8633-log-httpd\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.351184 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0baa8e1a-723d-49e2-9f65-0d95550a8633-run-httpd\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.355216 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-config-data\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.355341 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.355677 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.355951 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-scripts\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.358636 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0baa8e1a-723d-49e2-9f65-0d95550a8633-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.359095 5025 scope.go:117] "RemoveContainer" containerID="665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.359521 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5\": container with ID starting with 665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5 not found: ID does not exist" containerID="665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.359546 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5"} err="failed to get container status \"665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5\": rpc error: code = NotFound desc = could not find container \"665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5\": container with ID starting with 665b59acd2c6211c2fff2a5409187dd6bb4ee8cf4c3bb95caf9d112afcebb4d5 not found: ID does not exist" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.359564 5025 scope.go:117] "RemoveContainer" containerID="54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651" Oct 04 10:54:57 crc kubenswrapper[5025]: E1004 10:54:57.359870 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651\": container with ID starting with 54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651 not found: ID does not exist" containerID="54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.359918 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651"} err="failed to get container status \"54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651\": rpc error: code = NotFound desc = could not find container \"54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651\": container with ID starting with 54f61b3f61d676ea4d04ffbecc51e67f65af20e33865dd037e001c7487289651 not found: ID does not exist" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.370690 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz6wg\" (UniqueName: \"kubernetes.io/projected/0baa8e1a-723d-49e2-9f65-0d95550a8633-kube-api-access-nz6wg\") pod \"ceilometer-0\" (UID: \"0baa8e1a-723d-49e2-9f65-0d95550a8633\") " pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.451986 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s52n7\" (UniqueName: \"kubernetes.io/projected/adec9469-e051-42e1-a46c-87ce7f54aab3-kube-api-access-s52n7\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.452663 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-public-tls-certs\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.453383 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.453545 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adec9469-e051-42e1-a46c-87ce7f54aab3-logs\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.453698 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-config-data\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.453772 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.453989 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adec9469-e051-42e1-a46c-87ce7f54aab3-logs\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.456038 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-public-tls-certs\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.457529 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.458149 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-config-data\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.461578 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.470679 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s52n7\" (UniqueName: \"kubernetes.io/projected/adec9469-e051-42e1-a46c-87ce7f54aab3-kube-api-access-s52n7\") pod \"nova-api-0\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " pod="openstack/nova-api-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.622246 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 04 10:54:57 crc kubenswrapper[5025]: I1004 10:54:57.628748 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:54:58 crc kubenswrapper[5025]: I1004 10:54:58.125565 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 04 10:54:58 crc kubenswrapper[5025]: I1004 10:54:58.138477 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0baa8e1a-723d-49e2-9f65-0d95550a8633","Type":"ContainerStarted","Data":"656fef72ae1fdd5c703774b3f36beb119ccc00a4fc7c6135c52f52dc42514829"} Oct 04 10:54:58 crc kubenswrapper[5025]: I1004 10:54:58.253076 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:54:58 crc kubenswrapper[5025]: I1004 10:54:58.422060 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43ca7288-b2d5-4df8-9776-677efaf39087" path="/var/lib/kubelet/pods/43ca7288-b2d5-4df8-9776-677efaf39087/volumes" Oct 04 10:54:58 crc kubenswrapper[5025]: I1004 10:54:58.423053 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eee9436c-7c2b-4ccf-9218-1e2da6e96e6a" path="/var/lib/kubelet/pods/eee9436c-7c2b-4ccf-9218-1e2da6e96e6a/volumes" Oct 04 10:54:58 crc kubenswrapper[5025]: I1004 10:54:58.452301 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 10:54:58 crc kubenswrapper[5025]: I1004 10:54:58.452345 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 10:54:58 crc kubenswrapper[5025]: I1004 10:54:58.476570 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:58 crc kubenswrapper[5025]: I1004 10:54:58.520427 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.150899 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0baa8e1a-723d-49e2-9f65-0d95550a8633","Type":"ContainerStarted","Data":"8d7f700fe289bc3aa8c3bb375be12f1548ba5c1837222cc7af0ff5bac3884b87"} Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.153133 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adec9469-e051-42e1-a46c-87ce7f54aab3","Type":"ContainerStarted","Data":"29e9d50ccd3e83d63217623c97ff5a6fc362c1de005e9a19d50f8373368ab1b3"} Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.153176 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adec9469-e051-42e1-a46c-87ce7f54aab3","Type":"ContainerStarted","Data":"736c3219df1ef85213d8e33524f7aa40ff317e2c6d95580b6fd88acb6617cecb"} Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.153186 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adec9469-e051-42e1-a46c-87ce7f54aab3","Type":"ContainerStarted","Data":"ba386e158bf20fd26224782e7d9830c3e4c03ed1a02450f523c28aa377810790"} Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.176658 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.180952 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.180903153 podStartE2EDuration="2.180903153s" podCreationTimestamp="2025-10-04 10:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:54:59.176861242 +0000 UTC m=+1227.601828122" watchObservedRunningTime="2025-10-04 10:54:59.180903153 +0000 UTC m=+1227.605870033" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.470459 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.470754 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.496068 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-bs82b"] Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.497493 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.516988 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.517201 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.518957 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bs82b"] Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.605903 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.605971 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-config-data\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.606335 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-scripts\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.606424 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr4dg\" (UniqueName: \"kubernetes.io/projected/65baca16-d04e-44ea-8b1b-0f573bbd0d21-kube-api-access-qr4dg\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.707875 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.708205 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-config-data\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.708311 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-scripts\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.708347 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr4dg\" (UniqueName: \"kubernetes.io/projected/65baca16-d04e-44ea-8b1b-0f573bbd0d21-kube-api-access-qr4dg\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.715728 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.715792 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-config-data\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.716670 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-scripts\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.730526 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr4dg\" (UniqueName: \"kubernetes.io/projected/65baca16-d04e-44ea-8b1b-0f573bbd0d21-kube-api-access-qr4dg\") pod \"nova-cell1-cell-mapping-bs82b\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:54:59 crc kubenswrapper[5025]: I1004 10:54:59.828990 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.078421 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.147629 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-ngr4b"] Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.147871 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" podUID="6d8b4798-48b3-40d0-8330-2c7294991b3e" containerName="dnsmasq-dns" containerID="cri-o://fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35" gracePeriod=10 Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.186967 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0baa8e1a-723d-49e2-9f65-0d95550a8633","Type":"ContainerStarted","Data":"6db1dda2fe44dfc3e615b5a927e181d70ed415ec2b30ff782273afaf5e6b2ef3"} Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.344970 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bs82b"] Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.747423 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.835157 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2czh\" (UniqueName: \"kubernetes.io/projected/6d8b4798-48b3-40d0-8330-2c7294991b3e-kube-api-access-n2czh\") pod \"6d8b4798-48b3-40d0-8330-2c7294991b3e\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.835212 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-config\") pod \"6d8b4798-48b3-40d0-8330-2c7294991b3e\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.835267 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-nb\") pod \"6d8b4798-48b3-40d0-8330-2c7294991b3e\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.835311 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-svc\") pod \"6d8b4798-48b3-40d0-8330-2c7294991b3e\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.835349 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-swift-storage-0\") pod \"6d8b4798-48b3-40d0-8330-2c7294991b3e\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.835386 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-sb\") pod \"6d8b4798-48b3-40d0-8330-2c7294991b3e\" (UID: \"6d8b4798-48b3-40d0-8330-2c7294991b3e\") " Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.839874 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d8b4798-48b3-40d0-8330-2c7294991b3e-kube-api-access-n2czh" (OuterVolumeSpecName: "kube-api-access-n2czh") pod "6d8b4798-48b3-40d0-8330-2c7294991b3e" (UID: "6d8b4798-48b3-40d0-8330-2c7294991b3e"). InnerVolumeSpecName "kube-api-access-n2czh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.886782 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6d8b4798-48b3-40d0-8330-2c7294991b3e" (UID: "6d8b4798-48b3-40d0-8330-2c7294991b3e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.898820 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6d8b4798-48b3-40d0-8330-2c7294991b3e" (UID: "6d8b4798-48b3-40d0-8330-2c7294991b3e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.901657 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6d8b4798-48b3-40d0-8330-2c7294991b3e" (UID: "6d8b4798-48b3-40d0-8330-2c7294991b3e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.906476 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-config" (OuterVolumeSpecName: "config") pod "6d8b4798-48b3-40d0-8330-2c7294991b3e" (UID: "6d8b4798-48b3-40d0-8330-2c7294991b3e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.923562 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6d8b4798-48b3-40d0-8330-2c7294991b3e" (UID: "6d8b4798-48b3-40d0-8330-2c7294991b3e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.938092 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2czh\" (UniqueName: \"kubernetes.io/projected/6d8b4798-48b3-40d0-8330-2c7294991b3e-kube-api-access-n2czh\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.938139 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.938155 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.938169 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.938181 5025 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:00 crc kubenswrapper[5025]: I1004 10:55:00.938192 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6d8b4798-48b3-40d0-8330-2c7294991b3e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.204876 5025 generic.go:334] "Generic (PLEG): container finished" podID="6d8b4798-48b3-40d0-8330-2c7294991b3e" containerID="fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35" exitCode=0 Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.204936 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" event={"ID":"6d8b4798-48b3-40d0-8330-2c7294991b3e","Type":"ContainerDied","Data":"fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35"} Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.204961 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" event={"ID":"6d8b4798-48b3-40d0-8330-2c7294991b3e","Type":"ContainerDied","Data":"7abad6fce4504da7ec274c39bf669ed2e8be13d2725ea92e1279467096df2214"} Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.204977 5025 scope.go:117] "RemoveContainer" containerID="fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35" Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.205149 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-ngr4b" Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.208803 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bs82b" event={"ID":"65baca16-d04e-44ea-8b1b-0f573bbd0d21","Type":"ContainerStarted","Data":"b9838b5f5a4b67498f3df08464bb9d02bc5ce4e50e8ba8b921cf1702af0a1a3d"} Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.208865 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bs82b" event={"ID":"65baca16-d04e-44ea-8b1b-0f573bbd0d21","Type":"ContainerStarted","Data":"f1bad805e0549ea6996e42190d29caab6413fc9bd6d2fd75d6e0b66fb4bb380f"} Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.214881 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0baa8e1a-723d-49e2-9f65-0d95550a8633","Type":"ContainerStarted","Data":"9593e46dcc4727886be3d7b87688798b9adfe448fd316bb4a75e28cc79e24a9f"} Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.239652 5025 scope.go:117] "RemoveContainer" containerID="40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09" Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.241677 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-bs82b" podStartSLOduration=2.241657615 podStartE2EDuration="2.241657615s" podCreationTimestamp="2025-10-04 10:54:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:55:01.231733166 +0000 UTC m=+1229.656700046" watchObservedRunningTime="2025-10-04 10:55:01.241657615 +0000 UTC m=+1229.666624495" Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.264558 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-ngr4b"] Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.273144 5025 scope.go:117] "RemoveContainer" containerID="fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35" Oct 04 10:55:01 crc kubenswrapper[5025]: E1004 10:55:01.273592 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35\": container with ID starting with fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35 not found: ID does not exist" containerID="fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35" Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.273629 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35"} err="failed to get container status \"fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35\": rpc error: code = NotFound desc = could not find container \"fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35\": container with ID starting with fc1523f6117a0a9266bc768c057521cfd0b4c4b88a5c96e07d08231503e65c35 not found: ID does not exist" Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.273648 5025 scope.go:117] "RemoveContainer" containerID="40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09" Oct 04 10:55:01 crc kubenswrapper[5025]: E1004 10:55:01.274103 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09\": container with ID starting with 40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09 not found: ID does not exist" containerID="40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09" Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.274137 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09"} err="failed to get container status \"40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09\": rpc error: code = NotFound desc = could not find container \"40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09\": container with ID starting with 40aacd4e89c863904fee94c42664966721382ab699bbb584568d857793214a09 not found: ID does not exist" Oct 04 10:55:01 crc kubenswrapper[5025]: I1004 10:55:01.276191 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-ngr4b"] Oct 04 10:55:02 crc kubenswrapper[5025]: I1004 10:55:02.229034 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0baa8e1a-723d-49e2-9f65-0d95550a8633","Type":"ContainerStarted","Data":"e4512230c2ee442c40840d35ad37598efc7dd91af46dd1c873488ae74dc98939"} Oct 04 10:55:02 crc kubenswrapper[5025]: I1004 10:55:02.229542 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 04 10:55:02 crc kubenswrapper[5025]: I1004 10:55:02.254362 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.81991561 podStartE2EDuration="5.254341593s" podCreationTimestamp="2025-10-04 10:54:57 +0000 UTC" firstStartedPulling="2025-10-04 10:54:58.130850135 +0000 UTC m=+1226.555817015" lastFinishedPulling="2025-10-04 10:55:01.565276118 +0000 UTC m=+1229.990242998" observedRunningTime="2025-10-04 10:55:02.248584635 +0000 UTC m=+1230.673551515" watchObservedRunningTime="2025-10-04 10:55:02.254341593 +0000 UTC m=+1230.679308473" Oct 04 10:55:02 crc kubenswrapper[5025]: I1004 10:55:02.424707 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d8b4798-48b3-40d0-8330-2c7294991b3e" path="/var/lib/kubelet/pods/6d8b4798-48b3-40d0-8330-2c7294991b3e/volumes" Oct 04 10:55:06 crc kubenswrapper[5025]: I1004 10:55:06.268861 5025 generic.go:334] "Generic (PLEG): container finished" podID="65baca16-d04e-44ea-8b1b-0f573bbd0d21" containerID="b9838b5f5a4b67498f3df08464bb9d02bc5ce4e50e8ba8b921cf1702af0a1a3d" exitCode=0 Oct 04 10:55:06 crc kubenswrapper[5025]: I1004 10:55:06.268950 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bs82b" event={"ID":"65baca16-d04e-44ea-8b1b-0f573bbd0d21","Type":"ContainerDied","Data":"b9838b5f5a4b67498f3df08464bb9d02bc5ce4e50e8ba8b921cf1702af0a1a3d"} Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.631319 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.631379 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.741493 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.879107 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-scripts\") pod \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.879270 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-combined-ca-bundle\") pod \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.879439 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-config-data\") pod \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.879501 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr4dg\" (UniqueName: \"kubernetes.io/projected/65baca16-d04e-44ea-8b1b-0f573bbd0d21-kube-api-access-qr4dg\") pod \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\" (UID: \"65baca16-d04e-44ea-8b1b-0f573bbd0d21\") " Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.924034 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65baca16-d04e-44ea-8b1b-0f573bbd0d21-kube-api-access-qr4dg" (OuterVolumeSpecName: "kube-api-access-qr4dg") pod "65baca16-d04e-44ea-8b1b-0f573bbd0d21" (UID: "65baca16-d04e-44ea-8b1b-0f573bbd0d21"). InnerVolumeSpecName "kube-api-access-qr4dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.929977 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-scripts" (OuterVolumeSpecName: "scripts") pod "65baca16-d04e-44ea-8b1b-0f573bbd0d21" (UID: "65baca16-d04e-44ea-8b1b-0f573bbd0d21"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.933390 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-config-data" (OuterVolumeSpecName: "config-data") pod "65baca16-d04e-44ea-8b1b-0f573bbd0d21" (UID: "65baca16-d04e-44ea-8b1b-0f573bbd0d21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.965155 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65baca16-d04e-44ea-8b1b-0f573bbd0d21" (UID: "65baca16-d04e-44ea-8b1b-0f573bbd0d21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.982300 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.982366 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr4dg\" (UniqueName: \"kubernetes.io/projected/65baca16-d04e-44ea-8b1b-0f573bbd0d21-kube-api-access-qr4dg\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.982377 5025 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-scripts\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:07 crc kubenswrapper[5025]: I1004 10:55:07.982385 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65baca16-d04e-44ea-8b1b-0f573bbd0d21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.293715 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bs82b" event={"ID":"65baca16-d04e-44ea-8b1b-0f573bbd0d21","Type":"ContainerDied","Data":"f1bad805e0549ea6996e42190d29caab6413fc9bd6d2fd75d6e0b66fb4bb380f"} Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.293799 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1bad805e0549ea6996e42190d29caab6413fc9bd6d2fd75d6e0b66fb4bb380f" Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.293799 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bs82b" Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.487008 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.487383 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerName="nova-api-log" containerID="cri-o://736c3219df1ef85213d8e33524f7aa40ff317e2c6d95580b6fd88acb6617cecb" gracePeriod=30 Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.487525 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerName="nova-api-api" containerID="cri-o://29e9d50ccd3e83d63217623c97ff5a6fc362c1de005e9a19d50f8373368ab1b3" gracePeriod=30 Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.497207 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.499200 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.508750 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.513521 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e057802b-4b1c-493a-8ca1-15f81378a8fc" containerName="nova-scheduler-scheduler" containerID="cri-o://401857d48166fab6f580f3e80942cc354d7183f425fbe92d2b14f9b1a07390da" gracePeriod=30 Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.525353 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.525406 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": EOF" Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.539062 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": EOF" Oct 04 10:55:08 crc kubenswrapper[5025]: I1004 10:55:08.605003 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:55:09 crc kubenswrapper[5025]: E1004 10:55:09.248926 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="401857d48166fab6f580f3e80942cc354d7183f425fbe92d2b14f9b1a07390da" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 10:55:09 crc kubenswrapper[5025]: E1004 10:55:09.250865 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="401857d48166fab6f580f3e80942cc354d7183f425fbe92d2b14f9b1a07390da" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 10:55:09 crc kubenswrapper[5025]: E1004 10:55:09.252274 5025 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="401857d48166fab6f580f3e80942cc354d7183f425fbe92d2b14f9b1a07390da" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 04 10:55:09 crc kubenswrapper[5025]: E1004 10:55:09.252469 5025 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="e057802b-4b1c-493a-8ca1-15f81378a8fc" containerName="nova-scheduler-scheduler" Oct 04 10:55:09 crc kubenswrapper[5025]: I1004 10:55:09.303628 5025 generic.go:334] "Generic (PLEG): container finished" podID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerID="736c3219df1ef85213d8e33524f7aa40ff317e2c6d95580b6fd88acb6617cecb" exitCode=143 Oct 04 10:55:09 crc kubenswrapper[5025]: I1004 10:55:09.303707 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adec9469-e051-42e1-a46c-87ce7f54aab3","Type":"ContainerDied","Data":"736c3219df1ef85213d8e33524f7aa40ff317e2c6d95580b6fd88acb6617cecb"} Oct 04 10:55:09 crc kubenswrapper[5025]: I1004 10:55:09.311605 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 10:55:10 crc kubenswrapper[5025]: I1004 10:55:10.317563 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-log" containerID="cri-o://b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1" gracePeriod=30 Oct 04 10:55:10 crc kubenswrapper[5025]: I1004 10:55:10.317630 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-metadata" containerID="cri-o://f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694" gracePeriod=30 Oct 04 10:55:11 crc kubenswrapper[5025]: I1004 10:55:11.326924 5025 generic.go:334] "Generic (PLEG): container finished" podID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerID="b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1" exitCode=143 Oct 04 10:55:11 crc kubenswrapper[5025]: I1004 10:55:11.326962 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a195e5b2-84f4-4f36-9f73-0c0b30262c06","Type":"ContainerDied","Data":"b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1"} Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.350615 5025 generic.go:334] "Generic (PLEG): container finished" podID="e057802b-4b1c-493a-8ca1-15f81378a8fc" containerID="401857d48166fab6f580f3e80942cc354d7183f425fbe92d2b14f9b1a07390da" exitCode=0 Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.350670 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e057802b-4b1c-493a-8ca1-15f81378a8fc","Type":"ContainerDied","Data":"401857d48166fab6f580f3e80942cc354d7183f425fbe92d2b14f9b1a07390da"} Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.459162 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": dial tcp 10.217.0.194:8775: connect: connection refused" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.459297 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": dial tcp 10.217.0.194:8775: connect: connection refused" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.519929 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.687914 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdfrb\" (UniqueName: \"kubernetes.io/projected/e057802b-4b1c-493a-8ca1-15f81378a8fc-kube-api-access-kdfrb\") pod \"e057802b-4b1c-493a-8ca1-15f81378a8fc\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.688234 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-config-data\") pod \"e057802b-4b1c-493a-8ca1-15f81378a8fc\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.688265 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-combined-ca-bundle\") pod \"e057802b-4b1c-493a-8ca1-15f81378a8fc\" (UID: \"e057802b-4b1c-493a-8ca1-15f81378a8fc\") " Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.694404 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e057802b-4b1c-493a-8ca1-15f81378a8fc-kube-api-access-kdfrb" (OuterVolumeSpecName: "kube-api-access-kdfrb") pod "e057802b-4b1c-493a-8ca1-15f81378a8fc" (UID: "e057802b-4b1c-493a-8ca1-15f81378a8fc"). InnerVolumeSpecName "kube-api-access-kdfrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.725875 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e057802b-4b1c-493a-8ca1-15f81378a8fc" (UID: "e057802b-4b1c-493a-8ca1-15f81378a8fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.727909 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-config-data" (OuterVolumeSpecName: "config-data") pod "e057802b-4b1c-493a-8ca1-15f81378a8fc" (UID: "e057802b-4b1c-493a-8ca1-15f81378a8fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.789198 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.790502 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdfrb\" (UniqueName: \"kubernetes.io/projected/e057802b-4b1c-493a-8ca1-15f81378a8fc-kube-api-access-kdfrb\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.790525 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.790534 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e057802b-4b1c-493a-8ca1-15f81378a8fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.891925 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-combined-ca-bundle\") pod \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.891997 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-nova-metadata-tls-certs\") pod \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.892051 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a195e5b2-84f4-4f36-9f73-0c0b30262c06-logs\") pod \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.892092 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-config-data\") pod \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.892177 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkx6k\" (UniqueName: \"kubernetes.io/projected/a195e5b2-84f4-4f36-9f73-0c0b30262c06-kube-api-access-fkx6k\") pod \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\" (UID: \"a195e5b2-84f4-4f36-9f73-0c0b30262c06\") " Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.892818 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a195e5b2-84f4-4f36-9f73-0c0b30262c06-logs" (OuterVolumeSpecName: "logs") pod "a195e5b2-84f4-4f36-9f73-0c0b30262c06" (UID: "a195e5b2-84f4-4f36-9f73-0c0b30262c06"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.895814 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a195e5b2-84f4-4f36-9f73-0c0b30262c06-kube-api-access-fkx6k" (OuterVolumeSpecName: "kube-api-access-fkx6k") pod "a195e5b2-84f4-4f36-9f73-0c0b30262c06" (UID: "a195e5b2-84f4-4f36-9f73-0c0b30262c06"). InnerVolumeSpecName "kube-api-access-fkx6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.917854 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a195e5b2-84f4-4f36-9f73-0c0b30262c06" (UID: "a195e5b2-84f4-4f36-9f73-0c0b30262c06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.926403 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-config-data" (OuterVolumeSpecName: "config-data") pod "a195e5b2-84f4-4f36-9f73-0c0b30262c06" (UID: "a195e5b2-84f4-4f36-9f73-0c0b30262c06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.943312 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a195e5b2-84f4-4f36-9f73-0c0b30262c06" (UID: "a195e5b2-84f4-4f36-9f73-0c0b30262c06"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.994500 5025 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.994537 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a195e5b2-84f4-4f36-9f73-0c0b30262c06-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.994547 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.994556 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkx6k\" (UniqueName: \"kubernetes.io/projected/a195e5b2-84f4-4f36-9f73-0c0b30262c06-kube-api-access-fkx6k\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:13 crc kubenswrapper[5025]: I1004 10:55:13.994565 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a195e5b2-84f4-4f36-9f73-0c0b30262c06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.362761 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e057802b-4b1c-493a-8ca1-15f81378a8fc","Type":"ContainerDied","Data":"092a662ad4ec84e1d3d30131a517ef463ab15968bed94709a4076794f8c2fff9"} Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.362784 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.363133 5025 scope.go:117] "RemoveContainer" containerID="401857d48166fab6f580f3e80942cc354d7183f425fbe92d2b14f9b1a07390da" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.365092 5025 generic.go:334] "Generic (PLEG): container finished" podID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerID="f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694" exitCode=0 Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.365211 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.366108 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a195e5b2-84f4-4f36-9f73-0c0b30262c06","Type":"ContainerDied","Data":"f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694"} Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.366156 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a195e5b2-84f4-4f36-9f73-0c0b30262c06","Type":"ContainerDied","Data":"2f1e4cc45e5d2276b489a8b68727071b43c9f9e82d4370579ae06c2ef53ec282"} Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.367662 5025 generic.go:334] "Generic (PLEG): container finished" podID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerID="29e9d50ccd3e83d63217623c97ff5a6fc362c1de005e9a19d50f8373368ab1b3" exitCode=0 Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.367696 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adec9469-e051-42e1-a46c-87ce7f54aab3","Type":"ContainerDied","Data":"29e9d50ccd3e83d63217623c97ff5a6fc362c1de005e9a19d50f8373368ab1b3"} Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.367715 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"adec9469-e051-42e1-a46c-87ce7f54aab3","Type":"ContainerDied","Data":"ba386e158bf20fd26224782e7d9830c3e4c03ed1a02450f523c28aa377810790"} Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.367730 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba386e158bf20fd26224782e7d9830c3e4c03ed1a02450f523c28aa377810790" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.404777 5025 scope.go:117] "RemoveContainer" containerID="f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.407385 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.470884 5025 scope.go:117] "RemoveContainer" containerID="b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.515255 5025 scope.go:117] "RemoveContainer" containerID="f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.516230 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.516258 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.516275 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:55:14 crc kubenswrapper[5025]: E1004 10:55:14.518879 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694\": container with ID starting with f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694 not found: ID does not exist" containerID="f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.518918 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694"} err="failed to get container status \"f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694\": rpc error: code = NotFound desc = could not find container \"f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694\": container with ID starting with f211584b3ba40ea31797cb02ecd8a739e1f8b3c79388e55c3de9abc9af827694 not found: ID does not exist" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.518943 5025 scope.go:117] "RemoveContainer" containerID="b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1" Oct 04 10:55:14 crc kubenswrapper[5025]: E1004 10:55:14.518979 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65baca16-d04e-44ea-8b1b-0f573bbd0d21" containerName="nova-manage" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519006 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="65baca16-d04e-44ea-8b1b-0f573bbd0d21" containerName="nova-manage" Oct 04 10:55:14 crc kubenswrapper[5025]: E1004 10:55:14.519050 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerName="nova-api-api" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519058 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerName="nova-api-api" Oct 04 10:55:14 crc kubenswrapper[5025]: E1004 10:55:14.519083 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-metadata" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519091 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-metadata" Oct 04 10:55:14 crc kubenswrapper[5025]: E1004 10:55:14.519106 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-log" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519113 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-log" Oct 04 10:55:14 crc kubenswrapper[5025]: E1004 10:55:14.519133 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d8b4798-48b3-40d0-8330-2c7294991b3e" containerName="init" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519141 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d8b4798-48b3-40d0-8330-2c7294991b3e" containerName="init" Oct 04 10:55:14 crc kubenswrapper[5025]: E1004 10:55:14.519157 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerName="nova-api-log" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519164 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerName="nova-api-log" Oct 04 10:55:14 crc kubenswrapper[5025]: E1004 10:55:14.519178 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d8b4798-48b3-40d0-8330-2c7294991b3e" containerName="dnsmasq-dns" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519186 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d8b4798-48b3-40d0-8330-2c7294991b3e" containerName="dnsmasq-dns" Oct 04 10:55:14 crc kubenswrapper[5025]: E1004 10:55:14.519200 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e057802b-4b1c-493a-8ca1-15f81378a8fc" containerName="nova-scheduler-scheduler" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519209 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="e057802b-4b1c-493a-8ca1-15f81378a8fc" containerName="nova-scheduler-scheduler" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519407 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d8b4798-48b3-40d0-8330-2c7294991b3e" containerName="dnsmasq-dns" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519428 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerName="nova-api-api" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519435 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="e057802b-4b1c-493a-8ca1-15f81378a8fc" containerName="nova-scheduler-scheduler" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519446 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="65baca16-d04e-44ea-8b1b-0f573bbd0d21" containerName="nova-manage" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519454 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-metadata" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519467 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" containerName="nova-api-log" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.519480 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" containerName="nova-metadata-log" Oct 04 10:55:14 crc kubenswrapper[5025]: E1004 10:55:14.523137 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1\": container with ID starting with b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1 not found: ID does not exist" containerID="b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.523208 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1"} err="failed to get container status \"b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1\": rpc error: code = NotFound desc = could not find container \"b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1\": container with ID starting with b97988bd3922043fdc139c3ec9d748a8bc263e88fed0c88a6ab12b273fc8b2b1 not found: ID does not exist" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.523539 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.525199 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.528426 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.530088 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.555267 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.563208 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.572195 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.572315 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.573986 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.574547 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.610937 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adec9469-e051-42e1-a46c-87ce7f54aab3-logs\") pod \"adec9469-e051-42e1-a46c-87ce7f54aab3\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.611081 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s52n7\" (UniqueName: \"kubernetes.io/projected/adec9469-e051-42e1-a46c-87ce7f54aab3-kube-api-access-s52n7\") pod \"adec9469-e051-42e1-a46c-87ce7f54aab3\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.611190 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-config-data\") pod \"adec9469-e051-42e1-a46c-87ce7f54aab3\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.611285 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-public-tls-certs\") pod \"adec9469-e051-42e1-a46c-87ce7f54aab3\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.611313 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-internal-tls-certs\") pod \"adec9469-e051-42e1-a46c-87ce7f54aab3\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.611356 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-combined-ca-bundle\") pod \"adec9469-e051-42e1-a46c-87ce7f54aab3\" (UID: \"adec9469-e051-42e1-a46c-87ce7f54aab3\") " Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.611605 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8754265-dd19-4438-a891-94cf4a6ebe01-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e8754265-dd19-4438-a891-94cf4a6ebe01\") " pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.611643 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8754265-dd19-4438-a891-94cf4a6ebe01-config-data\") pod \"nova-scheduler-0\" (UID: \"e8754265-dd19-4438-a891-94cf4a6ebe01\") " pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.611664 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97nxp\" (UniqueName: \"kubernetes.io/projected/e8754265-dd19-4438-a891-94cf4a6ebe01-kube-api-access-97nxp\") pod \"nova-scheduler-0\" (UID: \"e8754265-dd19-4438-a891-94cf4a6ebe01\") " pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.611857 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adec9469-e051-42e1-a46c-87ce7f54aab3-logs" (OuterVolumeSpecName: "logs") pod "adec9469-e051-42e1-a46c-87ce7f54aab3" (UID: "adec9469-e051-42e1-a46c-87ce7f54aab3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.622835 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adec9469-e051-42e1-a46c-87ce7f54aab3-kube-api-access-s52n7" (OuterVolumeSpecName: "kube-api-access-s52n7") pod "adec9469-e051-42e1-a46c-87ce7f54aab3" (UID: "adec9469-e051-42e1-a46c-87ce7f54aab3"). InnerVolumeSpecName "kube-api-access-s52n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.638402 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "adec9469-e051-42e1-a46c-87ce7f54aab3" (UID: "adec9469-e051-42e1-a46c-87ce7f54aab3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.641851 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-config-data" (OuterVolumeSpecName: "config-data") pod "adec9469-e051-42e1-a46c-87ce7f54aab3" (UID: "adec9469-e051-42e1-a46c-87ce7f54aab3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.665501 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "adec9469-e051-42e1-a46c-87ce7f54aab3" (UID: "adec9469-e051-42e1-a46c-87ce7f54aab3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.676161 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "adec9469-e051-42e1-a46c-87ce7f54aab3" (UID: "adec9469-e051-42e1-a46c-87ce7f54aab3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713151 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f03a463d-561b-4517-9167-e2c66c8f323f-config-data\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713192 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f03a463d-561b-4517-9167-e2c66c8f323f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713259 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zmqz\" (UniqueName: \"kubernetes.io/projected/f03a463d-561b-4517-9167-e2c66c8f323f-kube-api-access-2zmqz\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713313 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8754265-dd19-4438-a891-94cf4a6ebe01-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e8754265-dd19-4438-a891-94cf4a6ebe01\") " pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713340 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8754265-dd19-4438-a891-94cf4a6ebe01-config-data\") pod \"nova-scheduler-0\" (UID: \"e8754265-dd19-4438-a891-94cf4a6ebe01\") " pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713355 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f03a463d-561b-4517-9167-e2c66c8f323f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713375 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97nxp\" (UniqueName: \"kubernetes.io/projected/e8754265-dd19-4438-a891-94cf4a6ebe01-kube-api-access-97nxp\") pod \"nova-scheduler-0\" (UID: \"e8754265-dd19-4438-a891-94cf4a6ebe01\") " pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713397 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f03a463d-561b-4517-9167-e2c66c8f323f-logs\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713442 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s52n7\" (UniqueName: \"kubernetes.io/projected/adec9469-e051-42e1-a46c-87ce7f54aab3-kube-api-access-s52n7\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713453 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713463 5025 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713473 5025 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713481 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adec9469-e051-42e1-a46c-87ce7f54aab3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.713490 5025 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/adec9469-e051-42e1-a46c-87ce7f54aab3-logs\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.716731 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8754265-dd19-4438-a891-94cf4a6ebe01-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e8754265-dd19-4438-a891-94cf4a6ebe01\") " pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.717224 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.717258 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.717290 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.717782 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b41afa99f31354e6a8480ad5274a7bc72b6f827a80b5595179f0a4d79cdf33b2"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.717869 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://b41afa99f31354e6a8480ad5274a7bc72b6f827a80b5595179f0a4d79cdf33b2" gracePeriod=600 Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.718512 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8754265-dd19-4438-a891-94cf4a6ebe01-config-data\") pod \"nova-scheduler-0\" (UID: \"e8754265-dd19-4438-a891-94cf4a6ebe01\") " pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.734403 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97nxp\" (UniqueName: \"kubernetes.io/projected/e8754265-dd19-4438-a891-94cf4a6ebe01-kube-api-access-97nxp\") pod \"nova-scheduler-0\" (UID: \"e8754265-dd19-4438-a891-94cf4a6ebe01\") " pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.815372 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f03a463d-561b-4517-9167-e2c66c8f323f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.815459 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f03a463d-561b-4517-9167-e2c66c8f323f-logs\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.815510 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f03a463d-561b-4517-9167-e2c66c8f323f-config-data\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.815534 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f03a463d-561b-4517-9167-e2c66c8f323f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.815624 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zmqz\" (UniqueName: \"kubernetes.io/projected/f03a463d-561b-4517-9167-e2c66c8f323f-kube-api-access-2zmqz\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.817875 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f03a463d-561b-4517-9167-e2c66c8f323f-logs\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.820079 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f03a463d-561b-4517-9167-e2c66c8f323f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.820370 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f03a463d-561b-4517-9167-e2c66c8f323f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.821390 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f03a463d-561b-4517-9167-e2c66c8f323f-config-data\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.833459 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zmqz\" (UniqueName: \"kubernetes.io/projected/f03a463d-561b-4517-9167-e2c66c8f323f-kube-api-access-2zmqz\") pod \"nova-metadata-0\" (UID: \"f03a463d-561b-4517-9167-e2c66c8f323f\") " pod="openstack/nova-metadata-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.848370 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 04 10:55:14 crc kubenswrapper[5025]: I1004 10:55:14.886981 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.382813 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="b41afa99f31354e6a8480ad5274a7bc72b6f827a80b5595179f0a4d79cdf33b2" exitCode=0 Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.382868 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"b41afa99f31354e6a8480ad5274a7bc72b6f827a80b5595179f0a4d79cdf33b2"} Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.382891 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.382905 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"56f39d1552d0ae55dd9141a0ac88b016c83fd194f8df926b4fe6a98c6db80007"} Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.382927 5025 scope.go:117] "RemoveContainer" containerID="1fa2c044718960ddbde296231834fb7f5a1519308e0e7742cc723c16081ff583" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.422527 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.428805 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.448958 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.452980 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.458664 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.460096 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.460901 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.461279 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.510578 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.531713 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.532054 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs98r\" (UniqueName: \"kubernetes.io/projected/9a3f816a-01d3-4823-b6e2-cf24e30a6735-kube-api-access-hs98r\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.532372 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-public-tls-certs\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.533175 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.534111 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-config-data\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.534639 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a3f816a-01d3-4823-b6e2-cf24e30a6735-logs\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: W1004 10:55:15.570918 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf03a463d_561b_4517_9167_e2c66c8f323f.slice/crio-8ee00b3eb110a85d209c4d5d9376e187343b83e03b4a7137e4cafd3beb0629d6 WatchSource:0}: Error finding container 8ee00b3eb110a85d209c4d5d9376e187343b83e03b4a7137e4cafd3beb0629d6: Status 404 returned error can't find the container with id 8ee00b3eb110a85d209c4d5d9376e187343b83e03b4a7137e4cafd3beb0629d6 Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.574788 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.636213 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-public-tls-certs\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.636429 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.637052 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-config-data\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.637340 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a3f816a-01d3-4823-b6e2-cf24e30a6735-logs\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.637429 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.637554 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs98r\" (UniqueName: \"kubernetes.io/projected/9a3f816a-01d3-4823-b6e2-cf24e30a6735-kube-api-access-hs98r\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.637666 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a3f816a-01d3-4823-b6e2-cf24e30a6735-logs\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.640782 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.640812 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-public-tls-certs\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.641417 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-config-data\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.641804 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3f816a-01d3-4823-b6e2-cf24e30a6735-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.655498 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs98r\" (UniqueName: \"kubernetes.io/projected/9a3f816a-01d3-4823-b6e2-cf24e30a6735-kube-api-access-hs98r\") pod \"nova-api-0\" (UID: \"9a3f816a-01d3-4823-b6e2-cf24e30a6735\") " pod="openstack/nova-api-0" Oct 04 10:55:15 crc kubenswrapper[5025]: I1004 10:55:15.784902 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.252644 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.397809 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f03a463d-561b-4517-9167-e2c66c8f323f","Type":"ContainerStarted","Data":"596f1fd55fb80c9fd39ab4691b11c0689a37735c8d540e8fe2e1762b7263729b"} Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.397868 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f03a463d-561b-4517-9167-e2c66c8f323f","Type":"ContainerStarted","Data":"063587ada72d404bd1e6e720a8d0fb5e40df1aee5e1aea948c37f51e1c37b63f"} Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.397889 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f03a463d-561b-4517-9167-e2c66c8f323f","Type":"ContainerStarted","Data":"8ee00b3eb110a85d209c4d5d9376e187343b83e03b4a7137e4cafd3beb0629d6"} Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.399585 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a3f816a-01d3-4823-b6e2-cf24e30a6735","Type":"ContainerStarted","Data":"efde97dfcac66f4fa2eed57954acb7e5e2d3844f642d052d23d0838dff3236d9"} Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.401235 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e8754265-dd19-4438-a891-94cf4a6ebe01","Type":"ContainerStarted","Data":"a3fc1eaf886746bef2c23eff7168ddd1b19b17ce245c630601c012aa8dda905f"} Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.401261 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e8754265-dd19-4438-a891-94cf4a6ebe01","Type":"ContainerStarted","Data":"eeabb0e9259976be91f300e682746985ba7a0166e7f22d013721a0dc07cb8ffe"} Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.421228 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.421207048 podStartE2EDuration="2.421207048s" podCreationTimestamp="2025-10-04 10:55:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:55:16.414825671 +0000 UTC m=+1244.839792551" watchObservedRunningTime="2025-10-04 10:55:16.421207048 +0000 UTC m=+1244.846173918" Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.421926 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a195e5b2-84f4-4f36-9f73-0c0b30262c06" path="/var/lib/kubelet/pods/a195e5b2-84f4-4f36-9f73-0c0b30262c06/volumes" Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.422676 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adec9469-e051-42e1-a46c-87ce7f54aab3" path="/var/lib/kubelet/pods/adec9469-e051-42e1-a46c-87ce7f54aab3/volumes" Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.423239 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e057802b-4b1c-493a-8ca1-15f81378a8fc" path="/var/lib/kubelet/pods/e057802b-4b1c-493a-8ca1-15f81378a8fc/volumes" Oct 04 10:55:16 crc kubenswrapper[5025]: I1004 10:55:16.436370 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.43635458 podStartE2EDuration="2.43635458s" podCreationTimestamp="2025-10-04 10:55:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:55:16.435038701 +0000 UTC m=+1244.860005581" watchObservedRunningTime="2025-10-04 10:55:16.43635458 +0000 UTC m=+1244.861321460" Oct 04 10:55:17 crc kubenswrapper[5025]: I1004 10:55:17.416819 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a3f816a-01d3-4823-b6e2-cf24e30a6735","Type":"ContainerStarted","Data":"ebb710124a590e312ab9dcf6a93bf58a32b157e3e592fa768e39080daf76659e"} Oct 04 10:55:17 crc kubenswrapper[5025]: I1004 10:55:17.417690 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9a3f816a-01d3-4823-b6e2-cf24e30a6735","Type":"ContainerStarted","Data":"72fbb2ac5baf06b1313665e6fde0f250c816f0ca5078d57b78b3d3404020ef38"} Oct 04 10:55:17 crc kubenswrapper[5025]: I1004 10:55:17.446780 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.446752102 podStartE2EDuration="2.446752102s" podCreationTimestamp="2025-10-04 10:55:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:55:17.440681264 +0000 UTC m=+1245.865648184" watchObservedRunningTime="2025-10-04 10:55:17.446752102 +0000 UTC m=+1245.871719022" Oct 04 10:55:19 crc kubenswrapper[5025]: I1004 10:55:19.849503 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 04 10:55:19 crc kubenswrapper[5025]: I1004 10:55:19.887978 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 10:55:19 crc kubenswrapper[5025]: I1004 10:55:19.888108 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 04 10:55:24 crc kubenswrapper[5025]: I1004 10:55:24.849533 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 04 10:55:24 crc kubenswrapper[5025]: I1004 10:55:24.875623 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 04 10:55:24 crc kubenswrapper[5025]: I1004 10:55:24.887579 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 10:55:24 crc kubenswrapper[5025]: I1004 10:55:24.887629 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 04 10:55:25 crc kubenswrapper[5025]: I1004 10:55:25.564155 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 04 10:55:25 crc kubenswrapper[5025]: I1004 10:55:25.786073 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 10:55:25 crc kubenswrapper[5025]: I1004 10:55:25.786133 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 04 10:55:25 crc kubenswrapper[5025]: I1004 10:55:25.907168 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f03a463d-561b-4517-9167-e2c66c8f323f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 10:55:25 crc kubenswrapper[5025]: I1004 10:55:25.907331 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f03a463d-561b-4517-9167-e2c66c8f323f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 10:55:26 crc kubenswrapper[5025]: I1004 10:55:26.798259 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9a3f816a-01d3-4823-b6e2-cf24e30a6735" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 10:55:26 crc kubenswrapper[5025]: I1004 10:55:26.798830 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9a3f816a-01d3-4823-b6e2-cf24e30a6735" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 04 10:55:27 crc kubenswrapper[5025]: I1004 10:55:27.630787 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 04 10:55:34 crc kubenswrapper[5025]: I1004 10:55:34.893511 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 10:55:34 crc kubenswrapper[5025]: I1004 10:55:34.894121 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 04 10:55:34 crc kubenswrapper[5025]: I1004 10:55:34.899243 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 10:55:34 crc kubenswrapper[5025]: I1004 10:55:34.907169 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 04 10:55:35 crc kubenswrapper[5025]: I1004 10:55:35.794248 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 10:55:35 crc kubenswrapper[5025]: I1004 10:55:35.794824 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 04 10:55:35 crc kubenswrapper[5025]: I1004 10:55:35.795080 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 10:55:35 crc kubenswrapper[5025]: I1004 10:55:35.795130 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 04 10:55:35 crc kubenswrapper[5025]: I1004 10:55:35.806006 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 10:55:35 crc kubenswrapper[5025]: I1004 10:55:35.806171 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 04 10:55:44 crc kubenswrapper[5025]: I1004 10:55:44.064564 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 10:55:44 crc kubenswrapper[5025]: I1004 10:55:44.907574 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 10:55:48 crc kubenswrapper[5025]: I1004 10:55:48.455080 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="a36a09c8-4254-4f14-bd39-7156cb462adb" containerName="rabbitmq" containerID="cri-o://2ab8fc08c3ee477353e833ec543df74b381fd508f89e05deb084314d2a145f33" gracePeriod=604796 Oct 04 10:55:49 crc kubenswrapper[5025]: I1004 10:55:49.207712 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" containerName="rabbitmq" containerID="cri-o://9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86" gracePeriod=604796 Oct 04 10:55:53 crc kubenswrapper[5025]: I1004 10:55:53.595242 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="a36a09c8-4254-4f14-bd39-7156cb462adb" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Oct 04 10:55:53 crc kubenswrapper[5025]: I1004 10:55:53.909438 5025 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 04 10:55:54 crc kubenswrapper[5025]: I1004 10:55:54.818753 5025 generic.go:334] "Generic (PLEG): container finished" podID="a36a09c8-4254-4f14-bd39-7156cb462adb" containerID="2ab8fc08c3ee477353e833ec543df74b381fd508f89e05deb084314d2a145f33" exitCode=0 Oct 04 10:55:54 crc kubenswrapper[5025]: I1004 10:55:54.819078 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a36a09c8-4254-4f14-bd39-7156cb462adb","Type":"ContainerDied","Data":"2ab8fc08c3ee477353e833ec543df74b381fd508f89e05deb084314d2a145f33"} Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.048781 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.208232 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-g8q66"] Oct 04 10:55:55 crc kubenswrapper[5025]: E1004 10:55:55.208667 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36a09c8-4254-4f14-bd39-7156cb462adb" containerName="rabbitmq" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.208682 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36a09c8-4254-4f14-bd39-7156cb462adb" containerName="rabbitmq" Oct 04 10:55:55 crc kubenswrapper[5025]: E1004 10:55:55.208705 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a36a09c8-4254-4f14-bd39-7156cb462adb" containerName="setup-container" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.208712 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="a36a09c8-4254-4f14-bd39-7156cb462adb" containerName="setup-container" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.208959 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="a36a09c8-4254-4f14-bd39-7156cb462adb" containerName="rabbitmq" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.210108 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.212193 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.225986 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-g8q66"] Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229560 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-confd\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229608 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-erlang-cookie\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229636 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229668 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-server-conf\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229703 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a36a09c8-4254-4f14-bd39-7156cb462adb-erlang-cookie-secret\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229732 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-config-data\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229774 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a36a09c8-4254-4f14-bd39-7156cb462adb-pod-info\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229849 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-tls\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229879 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-876lf\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-kube-api-access-876lf\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229941 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-plugins\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.229971 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-plugins-conf\") pod \"a36a09c8-4254-4f14-bd39-7156cb462adb\" (UID: \"a36a09c8-4254-4f14-bd39-7156cb462adb\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.230669 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.230886 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231141 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231181 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231229 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-config\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231252 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-svc\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231362 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231395 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hw8x\" (UniqueName: \"kubernetes.io/projected/5282b4d4-ee64-4924-8f06-78629add04d3-kube-api-access-8hw8x\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231429 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231494 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231619 5025 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231640 5025 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.231657 5025 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.240651 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-kube-api-access-876lf" (OuterVolumeSpecName: "kube-api-access-876lf") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "kube-api-access-876lf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.269323 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a36a09c8-4254-4f14-bd39-7156cb462adb-pod-info" (OuterVolumeSpecName: "pod-info") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.269323 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.269435 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.274115 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a36a09c8-4254-4f14-bd39-7156cb462adb-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.300749 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-config-data" (OuterVolumeSpecName: "config-data") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.319209 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-server-conf" (OuterVolumeSpecName: "server-conf") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.333634 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.333939 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.333989 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-config\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334007 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-svc\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334069 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334085 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hw8x\" (UniqueName: \"kubernetes.io/projected/5282b4d4-ee64-4924-8f06-78629add04d3-kube-api-access-8hw8x\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334107 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334171 5025 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a36a09c8-4254-4f14-bd39-7156cb462adb-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334182 5025 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334192 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-876lf\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-kube-api-access-876lf\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334212 5025 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334224 5025 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334232 5025 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a36a09c8-4254-4f14-bd39-7156cb462adb-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.334241 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a36a09c8-4254-4f14-bd39-7156cb462adb-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.335208 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.335776 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-svc\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.336308 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.336906 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-config\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.338374 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.344510 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.369495 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hw8x\" (UniqueName: \"kubernetes.io/projected/5282b4d4-ee64-4924-8f06-78629add04d3-kube-api-access-8hw8x\") pod \"dnsmasq-dns-5576978c7c-g8q66\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.387559 5025 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.411261 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a36a09c8-4254-4f14-bd39-7156cb462adb" (UID: "a36a09c8-4254-4f14-bd39-7156cb462adb"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.438321 5025 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a36a09c8-4254-4f14-bd39-7156cb462adb-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.438368 5025 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.530260 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.770916 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.838235 5025 generic.go:334] "Generic (PLEG): container finished" podID="9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" containerID="9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86" exitCode=0 Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.838321 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a","Type":"ContainerDied","Data":"9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86"} Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.838358 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a","Type":"ContainerDied","Data":"611be2be55fb857b34d266f1ee6ed53e3328c29a9e80563a5fd4dbcf3c7625b3"} Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.838382 5025 scope.go:117] "RemoveContainer" containerID="9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.838538 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.843504 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a36a09c8-4254-4f14-bd39-7156cb462adb","Type":"ContainerDied","Data":"b46770e10f8fe09a9e912f0790a414fc2630132ceaca368b643521a8d28b2687"} Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.843592 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.845646 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.845755 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-server-conf\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.845799 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-tls\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.845850 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfpqp\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-kube-api-access-kfpqp\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.845890 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-pod-info\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.845919 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-plugins-conf\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.845958 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-confd\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.846065 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-erlang-cookie\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.846099 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-config-data\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.846151 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-erlang-cookie-secret\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.846197 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-plugins\") pod \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\" (UID: \"9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a\") " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.847082 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.847553 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.847592 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.852729 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.853682 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.854216 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-pod-info" (OuterVolumeSpecName: "pod-info") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.855313 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-kube-api-access-kfpqp" (OuterVolumeSpecName: "kube-api-access-kfpqp") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "kube-api-access-kfpqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.869503 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.869530 5025 scope.go:117] "RemoveContainer" containerID="2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.887893 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-config-data" (OuterVolumeSpecName: "config-data") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.900597 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-server-conf" (OuterVolumeSpecName: "server-conf") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.948066 5025 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.948096 5025 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-server-conf\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.948107 5025 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.948118 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfpqp\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-kube-api-access-kfpqp\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.948127 5025 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-pod-info\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.948135 5025 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.948144 5025 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.948152 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.948160 5025 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.948168 5025 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.981423 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-g8q66"] Oct 04 10:55:55 crc kubenswrapper[5025]: I1004 10:55:55.988991 5025 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.004250 5025 scope.go:117] "RemoveContainer" containerID="9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86" Oct 04 10:55:56 crc kubenswrapper[5025]: E1004 10:55:56.006469 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86\": container with ID starting with 9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86 not found: ID does not exist" containerID="9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.006523 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86"} err="failed to get container status \"9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86\": rpc error: code = NotFound desc = could not find container \"9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86\": container with ID starting with 9a9dcc5dc69e799b4295f806fc94090946bc1688805dc5923e4e35b87b33ad86 not found: ID does not exist" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.006554 5025 scope.go:117] "RemoveContainer" containerID="2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea" Oct 04 10:55:56 crc kubenswrapper[5025]: E1004 10:55:56.007718 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea\": container with ID starting with 2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea not found: ID does not exist" containerID="2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.007758 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea"} err="failed to get container status \"2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea\": rpc error: code = NotFound desc = could not find container \"2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea\": container with ID starting with 2e388778a6e67d818a550c089cb958c62329fb986cf123abdf032e160d7329ea not found: ID does not exist" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.007784 5025 scope.go:117] "RemoveContainer" containerID="2ab8fc08c3ee477353e833ec543df74b381fd508f89e05deb084314d2a145f33" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.018151 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.031502 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.038388 5025 scope.go:117] "RemoveContainer" containerID="66815974cd0d2555f9c56c6b7046d5ea1aa08f08cf16377d7c60422b077b160a" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.039881 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" (UID: "9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.045237 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 10:55:56 crc kubenswrapper[5025]: E1004 10:55:56.045778 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" containerName="rabbitmq" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.045946 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" containerName="rabbitmq" Oct 04 10:55:56 crc kubenswrapper[5025]: E1004 10:55:56.046067 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" containerName="setup-container" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.046171 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" containerName="setup-container" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.046461 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" containerName="rabbitmq" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.053453 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.055252 5025 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.055303 5025 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.056305 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.057038 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2kgnj" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.057497 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.057608 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.057732 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.058156 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.058198 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.058236 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.156530 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.156585 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.156728 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.156760 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8adc9762-e9bc-427e-87a7-777fb153920c-config-data\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.156841 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8adc9762-e9bc-427e-87a7-777fb153920c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.156871 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8adc9762-e9bc-427e-87a7-777fb153920c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.156890 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8adc9762-e9bc-427e-87a7-777fb153920c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.156917 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8adc9762-e9bc-427e-87a7-777fb153920c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.156949 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7n7q\" (UniqueName: \"kubernetes.io/projected/8adc9762-e9bc-427e-87a7-777fb153920c-kube-api-access-k7n7q\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.156975 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.157083 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.183141 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.191385 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.212180 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.214121 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.218651 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.218898 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.219080 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.219236 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.223479 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.223905 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2vjs7" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.224191 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.229276 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.258869 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.258946 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.258988 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8adc9762-e9bc-427e-87a7-777fb153920c-config-data\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.259051 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8adc9762-e9bc-427e-87a7-777fb153920c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.259073 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8adc9762-e9bc-427e-87a7-777fb153920c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.259087 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8adc9762-e9bc-427e-87a7-777fb153920c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.259103 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8adc9762-e9bc-427e-87a7-777fb153920c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.259123 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7n7q\" (UniqueName: \"kubernetes.io/projected/8adc9762-e9bc-427e-87a7-777fb153920c-kube-api-access-k7n7q\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.259144 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.259172 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.259192 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.259606 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.260505 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8adc9762-e9bc-427e-87a7-777fb153920c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.260591 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.260763 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.261557 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8adc9762-e9bc-427e-87a7-777fb153920c-config-data\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.263118 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8adc9762-e9bc-427e-87a7-777fb153920c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.267073 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8adc9762-e9bc-427e-87a7-777fb153920c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.268939 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8adc9762-e9bc-427e-87a7-777fb153920c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.271122 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.274349 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8adc9762-e9bc-427e-87a7-777fb153920c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.278711 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7n7q\" (UniqueName: \"kubernetes.io/projected/8adc9762-e9bc-427e-87a7-777fb153920c-kube-api-access-k7n7q\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.325954 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"8adc9762-e9bc-427e-87a7-777fb153920c\") " pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.360950 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.361005 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ec47d2a-5987-4d71-bee2-648b3d664135-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.361081 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.361105 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ec47d2a-5987-4d71-bee2-648b3d664135-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.361133 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.361177 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.361218 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ec47d2a-5987-4d71-bee2-648b3d664135-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.361238 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ec47d2a-5987-4d71-bee2-648b3d664135-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.361276 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.361316 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtsbg\" (UniqueName: \"kubernetes.io/projected/0ec47d2a-5987-4d71-bee2-648b3d664135-kube-api-access-xtsbg\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.361374 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ec47d2a-5987-4d71-bee2-648b3d664135-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.375246 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.425174 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a" path="/var/lib/kubelet/pods/9a44ca98-3e64-4d0b-a7b6-8f4f1f67986a/volumes" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.425983 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a36a09c8-4254-4f14-bd39-7156cb462adb" path="/var/lib/kubelet/pods/a36a09c8-4254-4f14-bd39-7156cb462adb/volumes" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463263 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ec47d2a-5987-4d71-bee2-648b3d664135-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463332 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463359 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ec47d2a-5987-4d71-bee2-648b3d664135-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463409 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463436 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ec47d2a-5987-4d71-bee2-648b3d664135-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463458 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463497 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463571 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ec47d2a-5987-4d71-bee2-648b3d664135-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463594 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ec47d2a-5987-4d71-bee2-648b3d664135-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463633 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.463667 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtsbg\" (UniqueName: \"kubernetes.io/projected/0ec47d2a-5987-4d71-bee2-648b3d664135-kube-api-access-xtsbg\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.464852 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ec47d2a-5987-4d71-bee2-648b3d664135-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.465250 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.467883 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.471179 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ec47d2a-5987-4d71-bee2-648b3d664135-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.477871 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ec47d2a-5987-4d71-bee2-648b3d664135-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.477974 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ec47d2a-5987-4d71-bee2-648b3d664135-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.478105 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.479489 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ec47d2a-5987-4d71-bee2-648b3d664135-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.481485 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtsbg\" (UniqueName: \"kubernetes.io/projected/0ec47d2a-5987-4d71-bee2-648b3d664135-kube-api-access-xtsbg\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.482099 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.483394 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ec47d2a-5987-4d71-bee2-648b3d664135-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.556957 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ec47d2a-5987-4d71-bee2-648b3d664135\") " pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.847691 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.860741 5025 generic.go:334] "Generic (PLEG): container finished" podID="5282b4d4-ee64-4924-8f06-78629add04d3" containerID="5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80" exitCode=0 Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.860815 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" event={"ID":"5282b4d4-ee64-4924-8f06-78629add04d3","Type":"ContainerDied","Data":"5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80"} Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.860845 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" event={"ID":"5282b4d4-ee64-4924-8f06-78629add04d3","Type":"ContainerStarted","Data":"0698330ef7103a5ef1c89c60c175dc6e075998148acbf8eeecd0c1c30c772d58"} Oct 04 10:55:56 crc kubenswrapper[5025]: I1004 10:55:56.865889 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 04 10:55:56 crc kubenswrapper[5025]: W1004 10:55:56.894177 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8adc9762_e9bc_427e_87a7_777fb153920c.slice/crio-58bb2aa04b8f4d89b165d0600e33d0b1c6e2d9d7dc10135443f0ef98f64f8774 WatchSource:0}: Error finding container 58bb2aa04b8f4d89b165d0600e33d0b1c6e2d9d7dc10135443f0ef98f64f8774: Status 404 returned error can't find the container with id 58bb2aa04b8f4d89b165d0600e33d0b1c6e2d9d7dc10135443f0ef98f64f8774 Oct 04 10:55:57 crc kubenswrapper[5025]: I1004 10:55:57.373344 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 04 10:55:57 crc kubenswrapper[5025]: W1004 10:55:57.377860 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ec47d2a_5987_4d71_bee2_648b3d664135.slice/crio-bae4fceb17a895e490d79281a0545c58491617c8d19c0c2015da1502259bf010 WatchSource:0}: Error finding container bae4fceb17a895e490d79281a0545c58491617c8d19c0c2015da1502259bf010: Status 404 returned error can't find the container with id bae4fceb17a895e490d79281a0545c58491617c8d19c0c2015da1502259bf010 Oct 04 10:55:57 crc kubenswrapper[5025]: I1004 10:55:57.881544 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8adc9762-e9bc-427e-87a7-777fb153920c","Type":"ContainerStarted","Data":"58bb2aa04b8f4d89b165d0600e33d0b1c6e2d9d7dc10135443f0ef98f64f8774"} Oct 04 10:55:57 crc kubenswrapper[5025]: I1004 10:55:57.884721 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ec47d2a-5987-4d71-bee2-648b3d664135","Type":"ContainerStarted","Data":"bae4fceb17a895e490d79281a0545c58491617c8d19c0c2015da1502259bf010"} Oct 04 10:55:57 crc kubenswrapper[5025]: I1004 10:55:57.886807 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" event={"ID":"5282b4d4-ee64-4924-8f06-78629add04d3","Type":"ContainerStarted","Data":"663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca"} Oct 04 10:55:57 crc kubenswrapper[5025]: I1004 10:55:57.886964 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:55:57 crc kubenswrapper[5025]: I1004 10:55:57.914458 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" podStartSLOduration=2.914438246 podStartE2EDuration="2.914438246s" podCreationTimestamp="2025-10-04 10:55:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:55:57.9084117 +0000 UTC m=+1286.333378580" watchObservedRunningTime="2025-10-04 10:55:57.914438246 +0000 UTC m=+1286.339405126" Oct 04 10:55:58 crc kubenswrapper[5025]: I1004 10:55:58.906154 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ec47d2a-5987-4d71-bee2-648b3d664135","Type":"ContainerStarted","Data":"f1f6737d55a831ece569e01dd12e2231b3d73b3601aafcd9ef7cc1061f881f63"} Oct 04 10:55:58 crc kubenswrapper[5025]: I1004 10:55:58.911466 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8adc9762-e9bc-427e-87a7-777fb153920c","Type":"ContainerStarted","Data":"4ecf49708463c643fe22207af68eadddde36564649814b9f32961fb24e440879"} Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.533315 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.623538 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-glssw"] Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.623795 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" podUID="1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" containerName="dnsmasq-dns" containerID="cri-o://c46b538b3cb2c39640c9a6eb939dd6d80c04676def70e3fd0e24c4673f2c6bbb" gracePeriod=10 Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.814842 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-ts6zc"] Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.817073 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.843779 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-ts6zc"] Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.968117 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.968526 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-config\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.968554 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.968578 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.968632 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.968659 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:05 crc kubenswrapper[5025]: I1004 10:56:05.968687 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcs9q\" (UniqueName: \"kubernetes.io/projected/1529a91c-0604-4357-be99-3305358c188a-kube-api-access-vcs9q\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.023561 5025 generic.go:334] "Generic (PLEG): container finished" podID="1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" containerID="c46b538b3cb2c39640c9a6eb939dd6d80c04676def70e3fd0e24c4673f2c6bbb" exitCode=0 Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.023599 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" event={"ID":"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f","Type":"ContainerDied","Data":"c46b538b3cb2c39640c9a6eb939dd6d80c04676def70e3fd0e24c4673f2c6bbb"} Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.070780 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-config\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.070846 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.070882 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.070994 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.071054 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.071084 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcs9q\" (UniqueName: \"kubernetes.io/projected/1529a91c-0604-4357-be99-3305358c188a-kube-api-access-vcs9q\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.071246 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.072348 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.072917 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.073028 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.073146 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.073225 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-config\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.073770 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1529a91c-0604-4357-be99-3305358c188a-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.098147 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcs9q\" (UniqueName: \"kubernetes.io/projected/1529a91c-0604-4357-be99-3305358c188a-kube-api-access-vcs9q\") pod \"dnsmasq-dns-8c6f6df99-ts6zc\" (UID: \"1529a91c-0604-4357-be99-3305358c188a\") " pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.143369 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.237147 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.379960 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-nb\") pod \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.380026 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-sb\") pod \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.380051 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-svc\") pod \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.380080 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-swift-storage-0\") pod \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.380216 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7zgx\" (UniqueName: \"kubernetes.io/projected/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-kube-api-access-f7zgx\") pod \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.380300 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-config\") pod \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\" (UID: \"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f\") " Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.386615 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-kube-api-access-f7zgx" (OuterVolumeSpecName: "kube-api-access-f7zgx") pod "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" (UID: "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f"). InnerVolumeSpecName "kube-api-access-f7zgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.432782 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" (UID: "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.433033 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" (UID: "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.441924 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-config" (OuterVolumeSpecName: "config") pod "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" (UID: "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.443088 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" (UID: "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.449448 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" (UID: "1e897e29-fb1c-42ef-99d6-d1e601a5ab1f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.482659 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.482703 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.482720 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.482733 5025 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.482745 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7zgx\" (UniqueName: \"kubernetes.io/projected/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-kube-api-access-f7zgx\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.482758 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:06 crc kubenswrapper[5025]: I1004 10:56:06.585537 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-ts6zc"] Oct 04 10:56:07 crc kubenswrapper[5025]: I1004 10:56:07.037043 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" event={"ID":"1e897e29-fb1c-42ef-99d6-d1e601a5ab1f","Type":"ContainerDied","Data":"9d8e15c12f0b77325c1ff594615d3a87ea5b040e47ef6967dab636026ded6574"} Oct 04 10:56:07 crc kubenswrapper[5025]: I1004 10:56:07.037067 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-glssw" Oct 04 10:56:07 crc kubenswrapper[5025]: I1004 10:56:07.037741 5025 scope.go:117] "RemoveContainer" containerID="c46b538b3cb2c39640c9a6eb939dd6d80c04676def70e3fd0e24c4673f2c6bbb" Oct 04 10:56:07 crc kubenswrapper[5025]: I1004 10:56:07.038762 5025 generic.go:334] "Generic (PLEG): container finished" podID="1529a91c-0604-4357-be99-3305358c188a" containerID="4f424333d6d2e262f11e1709bc6fa2483c14c114aed7a8834249327e9e001ade" exitCode=0 Oct 04 10:56:07 crc kubenswrapper[5025]: I1004 10:56:07.038832 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" event={"ID":"1529a91c-0604-4357-be99-3305358c188a","Type":"ContainerDied","Data":"4f424333d6d2e262f11e1709bc6fa2483c14c114aed7a8834249327e9e001ade"} Oct 04 10:56:07 crc kubenswrapper[5025]: I1004 10:56:07.038884 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" event={"ID":"1529a91c-0604-4357-be99-3305358c188a","Type":"ContainerStarted","Data":"efa051504a234506800b03fc174aeee2b2d88c9dac35c920badd5e50339a060d"} Oct 04 10:56:07 crc kubenswrapper[5025]: I1004 10:56:07.061695 5025 scope.go:117] "RemoveContainer" containerID="fbf2ccaacec2c80fd6c1c7a5359b0484038b8df2161f7c15cf5192431f49bb6a" Oct 04 10:56:07 crc kubenswrapper[5025]: I1004 10:56:07.278562 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-glssw"] Oct 04 10:56:07 crc kubenswrapper[5025]: I1004 10:56:07.287477 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-glssw"] Oct 04 10:56:08 crc kubenswrapper[5025]: I1004 10:56:08.052080 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" event={"ID":"1529a91c-0604-4357-be99-3305358c188a","Type":"ContainerStarted","Data":"23464e92f40fbb3912f166e6bf653418134749e3459cb26f1764698270a15011"} Oct 04 10:56:08 crc kubenswrapper[5025]: I1004 10:56:08.053384 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:08 crc kubenswrapper[5025]: I1004 10:56:08.091621 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" podStartSLOduration=3.091603837 podStartE2EDuration="3.091603837s" podCreationTimestamp="2025-10-04 10:56:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:56:08.079931556 +0000 UTC m=+1296.504898436" watchObservedRunningTime="2025-10-04 10:56:08.091603837 +0000 UTC m=+1296.516570717" Oct 04 10:56:08 crc kubenswrapper[5025]: I1004 10:56:08.440241 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" path="/var/lib/kubelet/pods/1e897e29-fb1c-42ef-99d6-d1e601a5ab1f/volumes" Oct 04 10:56:16 crc kubenswrapper[5025]: I1004 10:56:16.145371 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8c6f6df99-ts6zc" Oct 04 10:56:16 crc kubenswrapper[5025]: I1004 10:56:16.231864 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-g8q66"] Oct 04 10:56:16 crc kubenswrapper[5025]: I1004 10:56:16.232278 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" podUID="5282b4d4-ee64-4924-8f06-78629add04d3" containerName="dnsmasq-dns" containerID="cri-o://663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca" gracePeriod=10 Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.729213 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.791348 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-config\") pod \"5282b4d4-ee64-4924-8f06-78629add04d3\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.791419 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-sb\") pod \"5282b4d4-ee64-4924-8f06-78629add04d3\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.791521 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-openstack-edpm-ipam\") pod \"5282b4d4-ee64-4924-8f06-78629add04d3\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.791569 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-nb\") pod \"5282b4d4-ee64-4924-8f06-78629add04d3\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.791600 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-swift-storage-0\") pod \"5282b4d4-ee64-4924-8f06-78629add04d3\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.791633 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hw8x\" (UniqueName: \"kubernetes.io/projected/5282b4d4-ee64-4924-8f06-78629add04d3-kube-api-access-8hw8x\") pod \"5282b4d4-ee64-4924-8f06-78629add04d3\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.791756 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-svc\") pod \"5282b4d4-ee64-4924-8f06-78629add04d3\" (UID: \"5282b4d4-ee64-4924-8f06-78629add04d3\") " Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.798059 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5282b4d4-ee64-4924-8f06-78629add04d3-kube-api-access-8hw8x" (OuterVolumeSpecName: "kube-api-access-8hw8x") pod "5282b4d4-ee64-4924-8f06-78629add04d3" (UID: "5282b4d4-ee64-4924-8f06-78629add04d3"). InnerVolumeSpecName "kube-api-access-8hw8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.835640 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "5282b4d4-ee64-4924-8f06-78629add04d3" (UID: "5282b4d4-ee64-4924-8f06-78629add04d3"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.841723 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5282b4d4-ee64-4924-8f06-78629add04d3" (UID: "5282b4d4-ee64-4924-8f06-78629add04d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.843112 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5282b4d4-ee64-4924-8f06-78629add04d3" (UID: "5282b4d4-ee64-4924-8f06-78629add04d3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.848172 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5282b4d4-ee64-4924-8f06-78629add04d3" (UID: "5282b4d4-ee64-4924-8f06-78629add04d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.848764 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-config" (OuterVolumeSpecName: "config") pod "5282b4d4-ee64-4924-8f06-78629add04d3" (UID: "5282b4d4-ee64-4924-8f06-78629add04d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.851764 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5282b4d4-ee64-4924-8f06-78629add04d3" (UID: "5282b4d4-ee64-4924-8f06-78629add04d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.894195 5025 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-config\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.894235 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.894248 5025 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.894262 5025 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.894274 5025 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.894286 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hw8x\" (UniqueName: \"kubernetes.io/projected/5282b4d4-ee64-4924-8f06-78629add04d3-kube-api-access-8hw8x\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:16.894297 5025 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5282b4d4-ee64-4924-8f06-78629add04d3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.166959 5025 generic.go:334] "Generic (PLEG): container finished" podID="5282b4d4-ee64-4924-8f06-78629add04d3" containerID="663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca" exitCode=0 Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.167035 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" event={"ID":"5282b4d4-ee64-4924-8f06-78629add04d3","Type":"ContainerDied","Data":"663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca"} Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.167068 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" event={"ID":"5282b4d4-ee64-4924-8f06-78629add04d3","Type":"ContainerDied","Data":"0698330ef7103a5ef1c89c60c175dc6e075998148acbf8eeecd0c1c30c772d58"} Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.167089 5025 scope.go:117] "RemoveContainer" containerID="663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.167089 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-g8q66" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.207658 5025 scope.go:117] "RemoveContainer" containerID="5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.213327 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-g8q66"] Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.222904 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-g8q66"] Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.232573 5025 scope.go:117] "RemoveContainer" containerID="663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca" Oct 04 10:56:17 crc kubenswrapper[5025]: E1004 10:56:17.233186 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca\": container with ID starting with 663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca not found: ID does not exist" containerID="663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.233235 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca"} err="failed to get container status \"663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca\": rpc error: code = NotFound desc = could not find container \"663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca\": container with ID starting with 663d63e81beb723c01ea1f8bb10bf75ddbcfc09a2cef45e2d7c5033fc4b218ca not found: ID does not exist" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.233266 5025 scope.go:117] "RemoveContainer" containerID="5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80" Oct 04 10:56:17 crc kubenswrapper[5025]: E1004 10:56:17.233666 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80\": container with ID starting with 5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80 not found: ID does not exist" containerID="5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80" Oct 04 10:56:17 crc kubenswrapper[5025]: I1004 10:56:17.233716 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80"} err="failed to get container status \"5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80\": rpc error: code = NotFound desc = could not find container \"5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80\": container with ID starting with 5d837e913b370730f4bebc4bc5e0438174496d663b412db8e3933079942e4a80 not found: ID does not exist" Oct 04 10:56:18 crc kubenswrapper[5025]: I1004 10:56:18.428347 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5282b4d4-ee64-4924-8f06-78629add04d3" path="/var/lib/kubelet/pods/5282b4d4-ee64-4924-8f06-78629add04d3/volumes" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.382004 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9"] Oct 04 10:56:29 crc kubenswrapper[5025]: E1004 10:56:29.382979 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" containerName="dnsmasq-dns" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.382995 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" containerName="dnsmasq-dns" Oct 04 10:56:29 crc kubenswrapper[5025]: E1004 10:56:29.383038 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5282b4d4-ee64-4924-8f06-78629add04d3" containerName="dnsmasq-dns" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.383047 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5282b4d4-ee64-4924-8f06-78629add04d3" containerName="dnsmasq-dns" Oct 04 10:56:29 crc kubenswrapper[5025]: E1004 10:56:29.383068 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5282b4d4-ee64-4924-8f06-78629add04d3" containerName="init" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.383077 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5282b4d4-ee64-4924-8f06-78629add04d3" containerName="init" Oct 04 10:56:29 crc kubenswrapper[5025]: E1004 10:56:29.383117 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" containerName="init" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.383126 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" containerName="init" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.383375 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e897e29-fb1c-42ef-99d6-d1e601a5ab1f" containerName="dnsmasq-dns" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.383392 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="5282b4d4-ee64-4924-8f06-78629add04d3" containerName="dnsmasq-dns" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.384247 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.388722 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.388722 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.388771 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.389131 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.401373 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9"] Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.461462 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7t55\" (UniqueName: \"kubernetes.io/projected/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-kube-api-access-d7t55\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.461518 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.462107 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.462280 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.564640 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.564831 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.564921 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.564962 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7t55\" (UniqueName: \"kubernetes.io/projected/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-kube-api-access-d7t55\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.574561 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.575283 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.575680 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.592145 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7t55\" (UniqueName: \"kubernetes.io/projected/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-kube-api-access-d7t55\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:29 crc kubenswrapper[5025]: I1004 10:56:29.711380 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:30 crc kubenswrapper[5025]: I1004 10:56:30.294499 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9"] Oct 04 10:56:30 crc kubenswrapper[5025]: I1004 10:56:30.296996 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 10:56:30 crc kubenswrapper[5025]: I1004 10:56:30.311363 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" event={"ID":"cc9937b2-f2f6-4e38-8c69-aaf75422be6a","Type":"ContainerStarted","Data":"27afcc07d95167706d16efae800803de79dd08ff0fabdcf95ea5d0c16b289ff6"} Oct 04 10:56:32 crc kubenswrapper[5025]: I1004 10:56:32.333331 5025 generic.go:334] "Generic (PLEG): container finished" podID="8adc9762-e9bc-427e-87a7-777fb153920c" containerID="4ecf49708463c643fe22207af68eadddde36564649814b9f32961fb24e440879" exitCode=0 Oct 04 10:56:32 crc kubenswrapper[5025]: I1004 10:56:32.333537 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8adc9762-e9bc-427e-87a7-777fb153920c","Type":"ContainerDied","Data":"4ecf49708463c643fe22207af68eadddde36564649814b9f32961fb24e440879"} Oct 04 10:56:32 crc kubenswrapper[5025]: I1004 10:56:32.337627 5025 generic.go:334] "Generic (PLEG): container finished" podID="0ec47d2a-5987-4d71-bee2-648b3d664135" containerID="f1f6737d55a831ece569e01dd12e2231b3d73b3601aafcd9ef7cc1061f881f63" exitCode=0 Oct 04 10:56:32 crc kubenswrapper[5025]: I1004 10:56:32.337746 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ec47d2a-5987-4d71-bee2-648b3d664135","Type":"ContainerDied","Data":"f1f6737d55a831ece569e01dd12e2231b3d73b3601aafcd9ef7cc1061f881f63"} Oct 04 10:56:33 crc kubenswrapper[5025]: I1004 10:56:33.349769 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8adc9762-e9bc-427e-87a7-777fb153920c","Type":"ContainerStarted","Data":"568072cd6eadc3a20ffe1cb7cf9ae56041580962da61cc6dbbedbf186558ff45"} Oct 04 10:56:33 crc kubenswrapper[5025]: I1004 10:56:33.350394 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 04 10:56:33 crc kubenswrapper[5025]: I1004 10:56:33.352387 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ec47d2a-5987-4d71-bee2-648b3d664135","Type":"ContainerStarted","Data":"7beb965683879d04822d6366c1db3fdb7046fbece58ea06b93703caee6eb102d"} Oct 04 10:56:33 crc kubenswrapper[5025]: I1004 10:56:33.352564 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:56:33 crc kubenswrapper[5025]: I1004 10:56:33.375083 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.375063064 podStartE2EDuration="37.375063064s" podCreationTimestamp="2025-10-04 10:55:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:56:33.372024075 +0000 UTC m=+1321.796990955" watchObservedRunningTime="2025-10-04 10:56:33.375063064 +0000 UTC m=+1321.800029944" Oct 04 10:56:33 crc kubenswrapper[5025]: I1004 10:56:33.407684 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.407666625 podStartE2EDuration="37.407666625s" podCreationTimestamp="2025-10-04 10:55:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 10:56:33.401860416 +0000 UTC m=+1321.826827316" watchObservedRunningTime="2025-10-04 10:56:33.407666625 +0000 UTC m=+1321.832633505" Oct 04 10:56:38 crc kubenswrapper[5025]: I1004 10:56:38.874563 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 10:56:39 crc kubenswrapper[5025]: I1004 10:56:39.417462 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" event={"ID":"cc9937b2-f2f6-4e38-8c69-aaf75422be6a","Type":"ContainerStarted","Data":"c445c3c3f631953836e0f18940148bad7b912601c1c1bad3e49d808b3f3952a2"} Oct 04 10:56:39 crc kubenswrapper[5025]: I1004 10:56:39.454279 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" podStartSLOduration=1.87956962 podStartE2EDuration="10.454248413s" podCreationTimestamp="2025-10-04 10:56:29 +0000 UTC" firstStartedPulling="2025-10-04 10:56:30.296754244 +0000 UTC m=+1318.721721134" lastFinishedPulling="2025-10-04 10:56:38.871433037 +0000 UTC m=+1327.296399927" observedRunningTime="2025-10-04 10:56:39.4390528 +0000 UTC m=+1327.864019680" watchObservedRunningTime="2025-10-04 10:56:39.454248413 +0000 UTC m=+1327.879215333" Oct 04 10:56:46 crc kubenswrapper[5025]: I1004 10:56:46.379451 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 04 10:56:46 crc kubenswrapper[5025]: I1004 10:56:46.851231 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 04 10:56:50 crc kubenswrapper[5025]: I1004 10:56:50.567789 5025 generic.go:334] "Generic (PLEG): container finished" podID="cc9937b2-f2f6-4e38-8c69-aaf75422be6a" containerID="c445c3c3f631953836e0f18940148bad7b912601c1c1bad3e49d808b3f3952a2" exitCode=0 Oct 04 10:56:50 crc kubenswrapper[5025]: I1004 10:56:50.567912 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" event={"ID":"cc9937b2-f2f6-4e38-8c69-aaf75422be6a","Type":"ContainerDied","Data":"c445c3c3f631953836e0f18940148bad7b912601c1c1bad3e49d808b3f3952a2"} Oct 04 10:56:51 crc kubenswrapper[5025]: I1004 10:56:51.989939 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.121910 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7t55\" (UniqueName: \"kubernetes.io/projected/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-kube-api-access-d7t55\") pod \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.122080 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-repo-setup-combined-ca-bundle\") pod \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.122286 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-inventory\") pod \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.122310 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-ssh-key\") pod \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\" (UID: \"cc9937b2-f2f6-4e38-8c69-aaf75422be6a\") " Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.128054 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "cc9937b2-f2f6-4e38-8c69-aaf75422be6a" (UID: "cc9937b2-f2f6-4e38-8c69-aaf75422be6a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.128901 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-kube-api-access-d7t55" (OuterVolumeSpecName: "kube-api-access-d7t55") pod "cc9937b2-f2f6-4e38-8c69-aaf75422be6a" (UID: "cc9937b2-f2f6-4e38-8c69-aaf75422be6a"). InnerVolumeSpecName "kube-api-access-d7t55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.156034 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-inventory" (OuterVolumeSpecName: "inventory") pod "cc9937b2-f2f6-4e38-8c69-aaf75422be6a" (UID: "cc9937b2-f2f6-4e38-8c69-aaf75422be6a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.175389 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cc9937b2-f2f6-4e38-8c69-aaf75422be6a" (UID: "cc9937b2-f2f6-4e38-8c69-aaf75422be6a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.224455 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7t55\" (UniqueName: \"kubernetes.io/projected/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-kube-api-access-d7t55\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.224492 5025 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.224504 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.224514 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc9937b2-f2f6-4e38-8c69-aaf75422be6a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.597062 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" event={"ID":"cc9937b2-f2f6-4e38-8c69-aaf75422be6a","Type":"ContainerDied","Data":"27afcc07d95167706d16efae800803de79dd08ff0fabdcf95ea5d0c16b289ff6"} Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.597099 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27afcc07d95167706d16efae800803de79dd08ff0fabdcf95ea5d0c16b289ff6" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.597146 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.683452 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7"] Oct 04 10:56:52 crc kubenswrapper[5025]: E1004 10:56:52.683925 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc9937b2-f2f6-4e38-8c69-aaf75422be6a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.683947 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc9937b2-f2f6-4e38-8c69-aaf75422be6a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.684238 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc9937b2-f2f6-4e38-8c69-aaf75422be6a" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.685336 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.688791 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.689057 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.689366 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.689863 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.699577 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7"] Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.837947 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2xpd7\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.838004 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwnh7\" (UniqueName: \"kubernetes.io/projected/663838d1-caa8-481f-b9fe-b8f51fc6eed9-kube-api-access-zwnh7\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2xpd7\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.838345 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2xpd7\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.940262 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2xpd7\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.940365 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2xpd7\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.940394 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwnh7\" (UniqueName: \"kubernetes.io/projected/663838d1-caa8-481f-b9fe-b8f51fc6eed9-kube-api-access-zwnh7\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2xpd7\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.945419 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2xpd7\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.945794 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2xpd7\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:52 crc kubenswrapper[5025]: I1004 10:56:52.960693 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwnh7\" (UniqueName: \"kubernetes.io/projected/663838d1-caa8-481f-b9fe-b8f51fc6eed9-kube-api-access-zwnh7\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-2xpd7\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:53 crc kubenswrapper[5025]: I1004 10:56:53.020326 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:53 crc kubenswrapper[5025]: I1004 10:56:53.608172 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7"] Oct 04 10:56:53 crc kubenswrapper[5025]: W1004 10:56:53.611611 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod663838d1_caa8_481f_b9fe_b8f51fc6eed9.slice/crio-9442ec45049dc85e3e9468cff1172f3e8a4f26bee257c08cdc7f4d3d8db37656 WatchSource:0}: Error finding container 9442ec45049dc85e3e9468cff1172f3e8a4f26bee257c08cdc7f4d3d8db37656: Status 404 returned error can't find the container with id 9442ec45049dc85e3e9468cff1172f3e8a4f26bee257c08cdc7f4d3d8db37656 Oct 04 10:56:54 crc kubenswrapper[5025]: I1004 10:56:54.619698 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" event={"ID":"663838d1-caa8-481f-b9fe-b8f51fc6eed9","Type":"ContainerStarted","Data":"902744032a91bf7f3e40f6faab5e495298fd5de687cb9826626f27778f14380b"} Oct 04 10:56:54 crc kubenswrapper[5025]: I1004 10:56:54.620083 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" event={"ID":"663838d1-caa8-481f-b9fe-b8f51fc6eed9","Type":"ContainerStarted","Data":"9442ec45049dc85e3e9468cff1172f3e8a4f26bee257c08cdc7f4d3d8db37656"} Oct 04 10:56:54 crc kubenswrapper[5025]: I1004 10:56:54.651217 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" podStartSLOduration=2.24217433 podStartE2EDuration="2.651199264s" podCreationTimestamp="2025-10-04 10:56:52 +0000 UTC" firstStartedPulling="2025-10-04 10:56:53.614539596 +0000 UTC m=+1342.039506466" lastFinishedPulling="2025-10-04 10:56:54.02356451 +0000 UTC m=+1342.448531400" observedRunningTime="2025-10-04 10:56:54.64351176 +0000 UTC m=+1343.068478640" watchObservedRunningTime="2025-10-04 10:56:54.651199264 +0000 UTC m=+1343.076166144" Oct 04 10:56:57 crc kubenswrapper[5025]: I1004 10:56:57.658956 5025 generic.go:334] "Generic (PLEG): container finished" podID="663838d1-caa8-481f-b9fe-b8f51fc6eed9" containerID="902744032a91bf7f3e40f6faab5e495298fd5de687cb9826626f27778f14380b" exitCode=0 Oct 04 10:56:57 crc kubenswrapper[5025]: I1004 10:56:57.659113 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" event={"ID":"663838d1-caa8-481f-b9fe-b8f51fc6eed9","Type":"ContainerDied","Data":"902744032a91bf7f3e40f6faab5e495298fd5de687cb9826626f27778f14380b"} Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.049186 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.166863 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwnh7\" (UniqueName: \"kubernetes.io/projected/663838d1-caa8-481f-b9fe-b8f51fc6eed9-kube-api-access-zwnh7\") pod \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.167057 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-ssh-key\") pod \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.167092 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-inventory\") pod \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\" (UID: \"663838d1-caa8-481f-b9fe-b8f51fc6eed9\") " Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.179282 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/663838d1-caa8-481f-b9fe-b8f51fc6eed9-kube-api-access-zwnh7" (OuterVolumeSpecName: "kube-api-access-zwnh7") pod "663838d1-caa8-481f-b9fe-b8f51fc6eed9" (UID: "663838d1-caa8-481f-b9fe-b8f51fc6eed9"). InnerVolumeSpecName "kube-api-access-zwnh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.213495 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "663838d1-caa8-481f-b9fe-b8f51fc6eed9" (UID: "663838d1-caa8-481f-b9fe-b8f51fc6eed9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.221492 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-inventory" (OuterVolumeSpecName: "inventory") pod "663838d1-caa8-481f-b9fe-b8f51fc6eed9" (UID: "663838d1-caa8-481f-b9fe-b8f51fc6eed9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.269905 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwnh7\" (UniqueName: \"kubernetes.io/projected/663838d1-caa8-481f-b9fe-b8f51fc6eed9-kube-api-access-zwnh7\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.269944 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.269960 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/663838d1-caa8-481f-b9fe-b8f51fc6eed9-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.679323 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" event={"ID":"663838d1-caa8-481f-b9fe-b8f51fc6eed9","Type":"ContainerDied","Data":"9442ec45049dc85e3e9468cff1172f3e8a4f26bee257c08cdc7f4d3d8db37656"} Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.679363 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9442ec45049dc85e3e9468cff1172f3e8a4f26bee257c08cdc7f4d3d8db37656" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.679445 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-2xpd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.764235 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7"] Oct 04 10:56:59 crc kubenswrapper[5025]: E1004 10:56:59.765049 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="663838d1-caa8-481f-b9fe-b8f51fc6eed9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.765081 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="663838d1-caa8-481f-b9fe-b8f51fc6eed9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.765371 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="663838d1-caa8-481f-b9fe-b8f51fc6eed9" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.766170 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.768001 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.768746 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.768881 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.768995 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.784457 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7"] Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.888510 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.888575 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.888637 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8jhr\" (UniqueName: \"kubernetes.io/projected/28e59cf5-a704-4eb4-94e7-84a90ceca38d-kube-api-access-m8jhr\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.888657 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.990852 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.990946 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.991063 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8jhr\" (UniqueName: \"kubernetes.io/projected/28e59cf5-a704-4eb4-94e7-84a90ceca38d-kube-api-access-m8jhr\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.991116 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.996105 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.996118 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:56:59 crc kubenswrapper[5025]: I1004 10:56:59.996935 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:57:00 crc kubenswrapper[5025]: I1004 10:57:00.010404 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8jhr\" (UniqueName: \"kubernetes.io/projected/28e59cf5-a704-4eb4-94e7-84a90ceca38d-kube-api-access-m8jhr\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:57:00 crc kubenswrapper[5025]: I1004 10:57:00.132278 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 10:57:00 crc kubenswrapper[5025]: I1004 10:57:00.679400 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7"] Oct 04 10:57:01 crc kubenswrapper[5025]: I1004 10:57:01.696692 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" event={"ID":"28e59cf5-a704-4eb4-94e7-84a90ceca38d","Type":"ContainerStarted","Data":"fc3ae3b41199f7fd57cf5a1867dcfe59b5935d4a5ccc366b98eb9dffc66bd77c"} Oct 04 10:57:01 crc kubenswrapper[5025]: I1004 10:57:01.697303 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" event={"ID":"28e59cf5-a704-4eb4-94e7-84a90ceca38d","Type":"ContainerStarted","Data":"3d4de37d49f9797749516cffad9ee251f299d460c677c045135c7535852b72a7"} Oct 04 10:57:01 crc kubenswrapper[5025]: I1004 10:57:01.712309 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" podStartSLOduration=2.333999216 podStartE2EDuration="2.712294725s" podCreationTimestamp="2025-10-04 10:56:59 +0000 UTC" firstStartedPulling="2025-10-04 10:57:00.683078543 +0000 UTC m=+1349.108045423" lastFinishedPulling="2025-10-04 10:57:01.061374052 +0000 UTC m=+1349.486340932" observedRunningTime="2025-10-04 10:57:01.709271206 +0000 UTC m=+1350.134238086" watchObservedRunningTime="2025-10-04 10:57:01.712294725 +0000 UTC m=+1350.137261605" Oct 04 10:57:44 crc kubenswrapper[5025]: I1004 10:57:44.714068 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:57:44 crc kubenswrapper[5025]: I1004 10:57:44.714751 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:57:46 crc kubenswrapper[5025]: I1004 10:57:46.598769 5025 scope.go:117] "RemoveContainer" containerID="fc3f2fdb59c5ed52cd622fdb5948a52993f2678899ea2a126225ec9f0b242589" Oct 04 10:57:46 crc kubenswrapper[5025]: I1004 10:57:46.636141 5025 scope.go:117] "RemoveContainer" containerID="4ad132bcb6db1791f620f0fb33be8a0e5e1299b62a58e7914f83bfc9d68d491e" Oct 04 10:58:14 crc kubenswrapper[5025]: I1004 10:58:14.714094 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:58:14 crc kubenswrapper[5025]: I1004 10:58:14.714879 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:58:32 crc kubenswrapper[5025]: I1004 10:58:32.912665 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4ctn4"] Oct 04 10:58:32 crc kubenswrapper[5025]: I1004 10:58:32.916632 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:32 crc kubenswrapper[5025]: I1004 10:58:32.932692 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4ctn4"] Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.010194 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-utilities\") pod \"redhat-operators-4ctn4\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.010691 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b28nj\" (UniqueName: \"kubernetes.io/projected/c67c1b56-dd21-4460-962b-96312bd0d0fd-kube-api-access-b28nj\") pod \"redhat-operators-4ctn4\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.010791 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-catalog-content\") pod \"redhat-operators-4ctn4\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.112546 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-utilities\") pod \"redhat-operators-4ctn4\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.112623 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b28nj\" (UniqueName: \"kubernetes.io/projected/c67c1b56-dd21-4460-962b-96312bd0d0fd-kube-api-access-b28nj\") pod \"redhat-operators-4ctn4\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.112670 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-catalog-content\") pod \"redhat-operators-4ctn4\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.113372 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-catalog-content\") pod \"redhat-operators-4ctn4\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.113385 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-utilities\") pod \"redhat-operators-4ctn4\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.137306 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b28nj\" (UniqueName: \"kubernetes.io/projected/c67c1b56-dd21-4460-962b-96312bd0d0fd-kube-api-access-b28nj\") pod \"redhat-operators-4ctn4\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.246475 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.693502 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4ctn4"] Oct 04 10:58:33 crc kubenswrapper[5025]: I1004 10:58:33.792086 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ctn4" event={"ID":"c67c1b56-dd21-4460-962b-96312bd0d0fd","Type":"ContainerStarted","Data":"06557e4f4018418a306cefe0e4540baec9397457c57b9e9ce418b466edbdf91a"} Oct 04 10:58:34 crc kubenswrapper[5025]: I1004 10:58:34.813155 5025 generic.go:334] "Generic (PLEG): container finished" podID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerID="65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70" exitCode=0 Oct 04 10:58:34 crc kubenswrapper[5025]: I1004 10:58:34.813298 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ctn4" event={"ID":"c67c1b56-dd21-4460-962b-96312bd0d0fd","Type":"ContainerDied","Data":"65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70"} Oct 04 10:58:36 crc kubenswrapper[5025]: I1004 10:58:36.843598 5025 generic.go:334] "Generic (PLEG): container finished" podID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerID="b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557" exitCode=0 Oct 04 10:58:36 crc kubenswrapper[5025]: I1004 10:58:36.843645 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ctn4" event={"ID":"c67c1b56-dd21-4460-962b-96312bd0d0fd","Type":"ContainerDied","Data":"b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557"} Oct 04 10:58:37 crc kubenswrapper[5025]: I1004 10:58:37.855672 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ctn4" event={"ID":"c67c1b56-dd21-4460-962b-96312bd0d0fd","Type":"ContainerStarted","Data":"6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11"} Oct 04 10:58:37 crc kubenswrapper[5025]: I1004 10:58:37.879619 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4ctn4" podStartSLOduration=3.459810406 podStartE2EDuration="5.879594061s" podCreationTimestamp="2025-10-04 10:58:32 +0000 UTC" firstStartedPulling="2025-10-04 10:58:34.816957706 +0000 UTC m=+1443.241924626" lastFinishedPulling="2025-10-04 10:58:37.236741401 +0000 UTC m=+1445.661708281" observedRunningTime="2025-10-04 10:58:37.871881575 +0000 UTC m=+1446.296848455" watchObservedRunningTime="2025-10-04 10:58:37.879594061 +0000 UTC m=+1446.304560941" Oct 04 10:58:43 crc kubenswrapper[5025]: I1004 10:58:43.246614 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:43 crc kubenswrapper[5025]: I1004 10:58:43.247184 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:43 crc kubenswrapper[5025]: I1004 10:58:43.306791 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:43 crc kubenswrapper[5025]: I1004 10:58:43.958696 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:44 crc kubenswrapper[5025]: I1004 10:58:44.009618 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4ctn4"] Oct 04 10:58:44 crc kubenswrapper[5025]: I1004 10:58:44.714297 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 10:58:44 crc kubenswrapper[5025]: I1004 10:58:44.714386 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 10:58:44 crc kubenswrapper[5025]: I1004 10:58:44.714449 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 10:58:44 crc kubenswrapper[5025]: I1004 10:58:44.715491 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"56f39d1552d0ae55dd9141a0ac88b016c83fd194f8df926b4fe6a98c6db80007"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 10:58:44 crc kubenswrapper[5025]: I1004 10:58:44.715592 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://56f39d1552d0ae55dd9141a0ac88b016c83fd194f8df926b4fe6a98c6db80007" gracePeriod=600 Oct 04 10:58:44 crc kubenswrapper[5025]: I1004 10:58:44.928153 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="56f39d1552d0ae55dd9141a0ac88b016c83fd194f8df926b4fe6a98c6db80007" exitCode=0 Oct 04 10:58:44 crc kubenswrapper[5025]: I1004 10:58:44.928299 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"56f39d1552d0ae55dd9141a0ac88b016c83fd194f8df926b4fe6a98c6db80007"} Oct 04 10:58:44 crc kubenswrapper[5025]: I1004 10:58:44.929454 5025 scope.go:117] "RemoveContainer" containerID="b41afa99f31354e6a8480ad5274a7bc72b6f827a80b5595179f0a4d79cdf33b2" Oct 04 10:58:45 crc kubenswrapper[5025]: I1004 10:58:45.941615 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634"} Oct 04 10:58:45 crc kubenswrapper[5025]: I1004 10:58:45.941757 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4ctn4" podUID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerName="registry-server" containerID="cri-o://6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11" gracePeriod=2 Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.394820 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.586890 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b28nj\" (UniqueName: \"kubernetes.io/projected/c67c1b56-dd21-4460-962b-96312bd0d0fd-kube-api-access-b28nj\") pod \"c67c1b56-dd21-4460-962b-96312bd0d0fd\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.587051 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-utilities\") pod \"c67c1b56-dd21-4460-962b-96312bd0d0fd\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.587147 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-catalog-content\") pod \"c67c1b56-dd21-4460-962b-96312bd0d0fd\" (UID: \"c67c1b56-dd21-4460-962b-96312bd0d0fd\") " Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.588283 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-utilities" (OuterVolumeSpecName: "utilities") pod "c67c1b56-dd21-4460-962b-96312bd0d0fd" (UID: "c67c1b56-dd21-4460-962b-96312bd0d0fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.588795 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.593424 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c67c1b56-dd21-4460-962b-96312bd0d0fd-kube-api-access-b28nj" (OuterVolumeSpecName: "kube-api-access-b28nj") pod "c67c1b56-dd21-4460-962b-96312bd0d0fd" (UID: "c67c1b56-dd21-4460-962b-96312bd0d0fd"). InnerVolumeSpecName "kube-api-access-b28nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.668452 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c67c1b56-dd21-4460-962b-96312bd0d0fd" (UID: "c67c1b56-dd21-4460-962b-96312bd0d0fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.690272 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c67c1b56-dd21-4460-962b-96312bd0d0fd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.690304 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b28nj\" (UniqueName: \"kubernetes.io/projected/c67c1b56-dd21-4460-962b-96312bd0d0fd-kube-api-access-b28nj\") on node \"crc\" DevicePath \"\"" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.752340 5025 scope.go:117] "RemoveContainer" containerID="64a255b2f77c316007bca29fbea4431bb259ddb5e8fe92d31a3ac65b6b471726" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.941844 5025 scope.go:117] "RemoveContainer" containerID="4146b8c8b42f82b144405ec2d0a537654909fd161ae073e045f001b05fa05cd0" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.959957 5025 generic.go:334] "Generic (PLEG): container finished" podID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerID="6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11" exitCode=0 Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.960329 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4ctn4" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.960313 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ctn4" event={"ID":"c67c1b56-dd21-4460-962b-96312bd0d0fd","Type":"ContainerDied","Data":"6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11"} Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.961179 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4ctn4" event={"ID":"c67c1b56-dd21-4460-962b-96312bd0d0fd","Type":"ContainerDied","Data":"06557e4f4018418a306cefe0e4540baec9397457c57b9e9ce418b466edbdf91a"} Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.961224 5025 scope.go:117] "RemoveContainer" containerID="6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.967947 5025 scope.go:117] "RemoveContainer" containerID="42c2292ecadfbc91dc0e9c87d4f8e32fcb6859f8c5de5d74f458cae466267bf1" Oct 04 10:58:46 crc kubenswrapper[5025]: I1004 10:58:46.990114 5025 scope.go:117] "RemoveContainer" containerID="b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.000917 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4ctn4"] Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.007132 5025 scope.go:117] "RemoveContainer" containerID="d6cafb614f4d4c7ccb64f8e00d276df76d61213552f09efaff5591b1e84d47d1" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.009901 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4ctn4"] Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.020618 5025 scope.go:117] "RemoveContainer" containerID="65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.061469 5025 scope.go:117] "RemoveContainer" containerID="6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11" Oct 04 10:58:47 crc kubenswrapper[5025]: E1004 10:58:47.062481 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11\": container with ID starting with 6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11 not found: ID does not exist" containerID="6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.062567 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11"} err="failed to get container status \"6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11\": rpc error: code = NotFound desc = could not find container \"6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11\": container with ID starting with 6a9c6bf03a629b8e87130390a0c3e13251a77ad47e261180341fe600d70cba11 not found: ID does not exist" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.062600 5025 scope.go:117] "RemoveContainer" containerID="b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557" Oct 04 10:58:47 crc kubenswrapper[5025]: E1004 10:58:47.063278 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557\": container with ID starting with b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557 not found: ID does not exist" containerID="b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.063324 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557"} err="failed to get container status \"b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557\": rpc error: code = NotFound desc = could not find container \"b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557\": container with ID starting with b1cb0c49607e06c4dd8ded0d775bd9db3ec4130857b279232ad789c1eea22557 not found: ID does not exist" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.063352 5025 scope.go:117] "RemoveContainer" containerID="65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70" Oct 04 10:58:47 crc kubenswrapper[5025]: E1004 10:58:47.063889 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70\": container with ID starting with 65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70 not found: ID does not exist" containerID="65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.063946 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70"} err="failed to get container status \"65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70\": rpc error: code = NotFound desc = could not find container \"65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70\": container with ID starting with 65c758b7d797d19c433e0aca1fef47ea97885aface83bc31c234e57b837c7b70 not found: ID does not exist" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.118205 5025 scope.go:117] "RemoveContainer" containerID="9599e792034366698ef8820b7faae2795782557237319bc10d68bba9acec0774" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.171633 5025 scope.go:117] "RemoveContainer" containerID="27287fca9695c5c44606f27eab7953333e63519a391fa405ee3566f719e2ee99" Oct 04 10:58:47 crc kubenswrapper[5025]: I1004 10:58:47.221315 5025 scope.go:117] "RemoveContainer" containerID="16fdc70e3b1d7f1e5bbf6f4cd085a28452869621c0a13d1016341a4f70694f64" Oct 04 10:58:48 crc kubenswrapper[5025]: I1004 10:58:48.430359 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c67c1b56-dd21-4460-962b-96312bd0d0fd" path="/var/lib/kubelet/pods/c67c1b56-dd21-4460-962b-96312bd0d0fd/volumes" Oct 04 10:58:54 crc kubenswrapper[5025]: I1004 10:58:54.947800 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9hb7p"] Oct 04 10:58:54 crc kubenswrapper[5025]: E1004 10:58:54.948923 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerName="extract-utilities" Oct 04 10:58:54 crc kubenswrapper[5025]: I1004 10:58:54.948943 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerName="extract-utilities" Oct 04 10:58:54 crc kubenswrapper[5025]: E1004 10:58:54.948971 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerName="extract-content" Oct 04 10:58:54 crc kubenswrapper[5025]: I1004 10:58:54.948978 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerName="extract-content" Oct 04 10:58:54 crc kubenswrapper[5025]: E1004 10:58:54.949003 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerName="registry-server" Oct 04 10:58:54 crc kubenswrapper[5025]: I1004 10:58:54.949029 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerName="registry-server" Oct 04 10:58:54 crc kubenswrapper[5025]: I1004 10:58:54.949264 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67c1b56-dd21-4460-962b-96312bd0d0fd" containerName="registry-server" Oct 04 10:58:54 crc kubenswrapper[5025]: I1004 10:58:54.950945 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:54 crc kubenswrapper[5025]: I1004 10:58:54.958124 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9hb7p"] Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.059777 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-utilities\") pod \"community-operators-9hb7p\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.060036 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7cvx\" (UniqueName: \"kubernetes.io/projected/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-kube-api-access-p7cvx\") pod \"community-operators-9hb7p\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.060073 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-catalog-content\") pod \"community-operators-9hb7p\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.161302 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7cvx\" (UniqueName: \"kubernetes.io/projected/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-kube-api-access-p7cvx\") pod \"community-operators-9hb7p\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.161368 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-catalog-content\") pod \"community-operators-9hb7p\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.161459 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-utilities\") pod \"community-operators-9hb7p\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.162118 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-utilities\") pod \"community-operators-9hb7p\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.162123 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-catalog-content\") pod \"community-operators-9hb7p\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.181110 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7cvx\" (UniqueName: \"kubernetes.io/projected/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-kube-api-access-p7cvx\") pod \"community-operators-9hb7p\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.269433 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:58:55 crc kubenswrapper[5025]: I1004 10:58:55.763091 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9hb7p"] Oct 04 10:58:56 crc kubenswrapper[5025]: I1004 10:58:56.060850 5025 generic.go:334] "Generic (PLEG): container finished" podID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerID="265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836" exitCode=0 Oct 04 10:58:56 crc kubenswrapper[5025]: I1004 10:58:56.060900 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb7p" event={"ID":"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2","Type":"ContainerDied","Data":"265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836"} Oct 04 10:58:56 crc kubenswrapper[5025]: I1004 10:58:56.060929 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb7p" event={"ID":"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2","Type":"ContainerStarted","Data":"f0f3407947f6f91ab052e94150e85aed3324a289af4a5fab4ece075a64b706c1"} Oct 04 10:58:57 crc kubenswrapper[5025]: I1004 10:58:57.070899 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb7p" event={"ID":"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2","Type":"ContainerStarted","Data":"9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73"} Oct 04 10:58:58 crc kubenswrapper[5025]: I1004 10:58:58.083030 5025 generic.go:334] "Generic (PLEG): container finished" podID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerID="9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73" exitCode=0 Oct 04 10:58:58 crc kubenswrapper[5025]: I1004 10:58:58.083150 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb7p" event={"ID":"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2","Type":"ContainerDied","Data":"9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73"} Oct 04 10:58:59 crc kubenswrapper[5025]: I1004 10:58:59.096818 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb7p" event={"ID":"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2","Type":"ContainerStarted","Data":"17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a"} Oct 04 10:58:59 crc kubenswrapper[5025]: I1004 10:58:59.145908 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9hb7p" podStartSLOduration=2.681660417 podStartE2EDuration="5.145874502s" podCreationTimestamp="2025-10-04 10:58:54 +0000 UTC" firstStartedPulling="2025-10-04 10:58:56.062355385 +0000 UTC m=+1464.487322265" lastFinishedPulling="2025-10-04 10:58:58.52656947 +0000 UTC m=+1466.951536350" observedRunningTime="2025-10-04 10:58:59.115879515 +0000 UTC m=+1467.540846405" watchObservedRunningTime="2025-10-04 10:58:59.145874502 +0000 UTC m=+1467.570841402" Oct 04 10:59:05 crc kubenswrapper[5025]: I1004 10:59:05.269576 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:59:05 crc kubenswrapper[5025]: I1004 10:59:05.270306 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:59:05 crc kubenswrapper[5025]: I1004 10:59:05.328173 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:59:06 crc kubenswrapper[5025]: I1004 10:59:06.224476 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:59:06 crc kubenswrapper[5025]: I1004 10:59:06.286467 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9hb7p"] Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.179374 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9hb7p" podUID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerName="registry-server" containerID="cri-o://17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a" gracePeriod=2 Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.647713 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.837948 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7cvx\" (UniqueName: \"kubernetes.io/projected/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-kube-api-access-p7cvx\") pod \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.838386 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-catalog-content\") pod \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.839333 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-utilities\") pod \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\" (UID: \"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2\") " Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.841316 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-utilities" (OuterVolumeSpecName: "utilities") pod "8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" (UID: "8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.847120 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-kube-api-access-p7cvx" (OuterVolumeSpecName: "kube-api-access-p7cvx") pod "8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" (UID: "8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2"). InnerVolumeSpecName "kube-api-access-p7cvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.892795 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" (UID: "8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.942492 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.942534 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 10:59:08 crc kubenswrapper[5025]: I1004 10:59:08.942545 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7cvx\" (UniqueName: \"kubernetes.io/projected/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2-kube-api-access-p7cvx\") on node \"crc\" DevicePath \"\"" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.193211 5025 generic.go:334] "Generic (PLEG): container finished" podID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerID="17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a" exitCode=0 Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.193361 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb7p" event={"ID":"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2","Type":"ContainerDied","Data":"17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a"} Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.193436 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hb7p" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.193573 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb7p" event={"ID":"8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2","Type":"ContainerDied","Data":"f0f3407947f6f91ab052e94150e85aed3324a289af4a5fab4ece075a64b706c1"} Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.193590 5025 scope.go:117] "RemoveContainer" containerID="17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.222672 5025 scope.go:117] "RemoveContainer" containerID="9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.247077 5025 scope.go:117] "RemoveContainer" containerID="265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.302438 5025 scope.go:117] "RemoveContainer" containerID="17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.304421 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9hb7p"] Oct 04 10:59:09 crc kubenswrapper[5025]: E1004 10:59:09.309568 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a\": container with ID starting with 17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a not found: ID does not exist" containerID="17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.309629 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a"} err="failed to get container status \"17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a\": rpc error: code = NotFound desc = could not find container \"17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a\": container with ID starting with 17c42c616003da57935aff507bae5edea01ebfcac0b65b8fcb4fd3e6bd81059a not found: ID does not exist" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.309670 5025 scope.go:117] "RemoveContainer" containerID="9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73" Oct 04 10:59:09 crc kubenswrapper[5025]: E1004 10:59:09.310276 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73\": container with ID starting with 9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73 not found: ID does not exist" containerID="9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.310325 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73"} err="failed to get container status \"9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73\": rpc error: code = NotFound desc = could not find container \"9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73\": container with ID starting with 9d8c41a55eaf8d3e08bcab43a1015a596bac207780fdd7fb39e36aa08458ef73 not found: ID does not exist" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.310371 5025 scope.go:117] "RemoveContainer" containerID="265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836" Oct 04 10:59:09 crc kubenswrapper[5025]: E1004 10:59:09.310705 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836\": container with ID starting with 265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836 not found: ID does not exist" containerID="265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.310750 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836"} err="failed to get container status \"265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836\": rpc error: code = NotFound desc = could not find container \"265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836\": container with ID starting with 265f1f17afe1b5dd530be37d3746328fb3a4101923b8c4f11f9f2f97a05e7836 not found: ID does not exist" Oct 04 10:59:09 crc kubenswrapper[5025]: I1004 10:59:09.325327 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9hb7p"] Oct 04 10:59:10 crc kubenswrapper[5025]: I1004 10:59:10.434294 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" path="/var/lib/kubelet/pods/8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2/volumes" Oct 04 10:59:58 crc kubenswrapper[5025]: I1004 10:59:58.838795 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vv75n"] Oct 04 10:59:58 crc kubenswrapper[5025]: E1004 10:59:58.839858 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerName="extract-utilities" Oct 04 10:59:58 crc kubenswrapper[5025]: I1004 10:59:58.839871 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerName="extract-utilities" Oct 04 10:59:58 crc kubenswrapper[5025]: E1004 10:59:58.839889 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerName="extract-content" Oct 04 10:59:58 crc kubenswrapper[5025]: I1004 10:59:58.839895 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerName="extract-content" Oct 04 10:59:58 crc kubenswrapper[5025]: E1004 10:59:58.839908 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerName="registry-server" Oct 04 10:59:58 crc kubenswrapper[5025]: I1004 10:59:58.839916 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerName="registry-server" Oct 04 10:59:58 crc kubenswrapper[5025]: I1004 10:59:58.840154 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5a2ce1-bb1e-4d07-a6b9-80c91baa83b2" containerName="registry-server" Oct 04 10:59:58 crc kubenswrapper[5025]: I1004 10:59:58.841808 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:58 crc kubenswrapper[5025]: I1004 10:59:58.853297 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vv75n"] Oct 04 10:59:58 crc kubenswrapper[5025]: I1004 10:59:58.950115 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-catalog-content\") pod \"certified-operators-vv75n\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:58 crc kubenswrapper[5025]: I1004 10:59:58.950217 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-utilities\") pod \"certified-operators-vv75n\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:58 crc kubenswrapper[5025]: I1004 10:59:58.950278 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp2ls\" (UniqueName: \"kubernetes.io/projected/d2f03f7c-2b0c-42f3-b2ec-121f95090044-kube-api-access-jp2ls\") pod \"certified-operators-vv75n\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.051447 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-catalog-content\") pod \"certified-operators-vv75n\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.051597 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-utilities\") pod \"certified-operators-vv75n\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.051651 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp2ls\" (UniqueName: \"kubernetes.io/projected/d2f03f7c-2b0c-42f3-b2ec-121f95090044-kube-api-access-jp2ls\") pod \"certified-operators-vv75n\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.051996 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-catalog-content\") pod \"certified-operators-vv75n\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.052191 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-utilities\") pod \"certified-operators-vv75n\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.084000 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp2ls\" (UniqueName: \"kubernetes.io/projected/d2f03f7c-2b0c-42f3-b2ec-121f95090044-kube-api-access-jp2ls\") pod \"certified-operators-vv75n\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.162471 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vv75n" Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.468990 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vv75n"] Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.744478 5025 generic.go:334] "Generic (PLEG): container finished" podID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerID="55557599490bd10653e5def755171c31b4e41d8b7c82eb997744578ce8272bb9" exitCode=0 Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.744514 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vv75n" event={"ID":"d2f03f7c-2b0c-42f3-b2ec-121f95090044","Type":"ContainerDied","Data":"55557599490bd10653e5def755171c31b4e41d8b7c82eb997744578ce8272bb9"} Oct 04 10:59:59 crc kubenswrapper[5025]: I1004 10:59:59.744537 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vv75n" event={"ID":"d2f03f7c-2b0c-42f3-b2ec-121f95090044","Type":"ContainerStarted","Data":"a9cbac56a4965f713d7d20e0c3096631e93106fb8aee0d01817f483b3a99edd0"} Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.142486 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq"] Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.143972 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.145993 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.152446 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq"] Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.174830 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.277726 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93845871-6c40-43cc-b144-257d6e4385af-config-volume\") pod \"collect-profiles-29326260-pmfcq\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.277776 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93845871-6c40-43cc-b144-257d6e4385af-secret-volume\") pod \"collect-profiles-29326260-pmfcq\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.277909 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljvkg\" (UniqueName: \"kubernetes.io/projected/93845871-6c40-43cc-b144-257d6e4385af-kube-api-access-ljvkg\") pod \"collect-profiles-29326260-pmfcq\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.379499 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93845871-6c40-43cc-b144-257d6e4385af-config-volume\") pod \"collect-profiles-29326260-pmfcq\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.379568 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93845871-6c40-43cc-b144-257d6e4385af-secret-volume\") pod \"collect-profiles-29326260-pmfcq\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.379687 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljvkg\" (UniqueName: \"kubernetes.io/projected/93845871-6c40-43cc-b144-257d6e4385af-kube-api-access-ljvkg\") pod \"collect-profiles-29326260-pmfcq\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.380723 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93845871-6c40-43cc-b144-257d6e4385af-config-volume\") pod \"collect-profiles-29326260-pmfcq\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.394310 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93845871-6c40-43cc-b144-257d6e4385af-secret-volume\") pod \"collect-profiles-29326260-pmfcq\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.398754 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljvkg\" (UniqueName: \"kubernetes.io/projected/93845871-6c40-43cc-b144-257d6e4385af-kube-api-access-ljvkg\") pod \"collect-profiles-29326260-pmfcq\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:00 crc kubenswrapper[5025]: I1004 11:00:00.493706 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:01 crc kubenswrapper[5025]: I1004 11:00:01.058639 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq"] Oct 04 11:00:01 crc kubenswrapper[5025]: I1004 11:00:01.766037 5025 generic.go:334] "Generic (PLEG): container finished" podID="93845871-6c40-43cc-b144-257d6e4385af" containerID="dbe57f783eb1381a271a6a4e58f90b594bf318a37cd0994a56eb904f52e7cc4d" exitCode=0 Oct 04 11:00:01 crc kubenswrapper[5025]: I1004 11:00:01.766220 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" event={"ID":"93845871-6c40-43cc-b144-257d6e4385af","Type":"ContainerDied","Data":"dbe57f783eb1381a271a6a4e58f90b594bf318a37cd0994a56eb904f52e7cc4d"} Oct 04 11:00:01 crc kubenswrapper[5025]: I1004 11:00:01.766483 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" event={"ID":"93845871-6c40-43cc-b144-257d6e4385af","Type":"ContainerStarted","Data":"03d08843008f6e0944870556cd80c87c4beb9fa3276db5872483d733e0a2f346"} Oct 04 11:00:01 crc kubenswrapper[5025]: I1004 11:00:01.769185 5025 generic.go:334] "Generic (PLEG): container finished" podID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerID="e039b4c7959039bf23c332d017663b7f9e2f31c714c062c63b034c891463e525" exitCode=0 Oct 04 11:00:01 crc kubenswrapper[5025]: I1004 11:00:01.769227 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vv75n" event={"ID":"d2f03f7c-2b0c-42f3-b2ec-121f95090044","Type":"ContainerDied","Data":"e039b4c7959039bf23c332d017663b7f9e2f31c714c062c63b034c891463e525"} Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.120192 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.239492 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljvkg\" (UniqueName: \"kubernetes.io/projected/93845871-6c40-43cc-b144-257d6e4385af-kube-api-access-ljvkg\") pod \"93845871-6c40-43cc-b144-257d6e4385af\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.239723 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93845871-6c40-43cc-b144-257d6e4385af-secret-volume\") pod \"93845871-6c40-43cc-b144-257d6e4385af\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.239746 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93845871-6c40-43cc-b144-257d6e4385af-config-volume\") pod \"93845871-6c40-43cc-b144-257d6e4385af\" (UID: \"93845871-6c40-43cc-b144-257d6e4385af\") " Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.240807 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93845871-6c40-43cc-b144-257d6e4385af-config-volume" (OuterVolumeSpecName: "config-volume") pod "93845871-6c40-43cc-b144-257d6e4385af" (UID: "93845871-6c40-43cc-b144-257d6e4385af"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.258817 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93845871-6c40-43cc-b144-257d6e4385af-kube-api-access-ljvkg" (OuterVolumeSpecName: "kube-api-access-ljvkg") pod "93845871-6c40-43cc-b144-257d6e4385af" (UID: "93845871-6c40-43cc-b144-257d6e4385af"). InnerVolumeSpecName "kube-api-access-ljvkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.258916 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93845871-6c40-43cc-b144-257d6e4385af-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "93845871-6c40-43cc-b144-257d6e4385af" (UID: "93845871-6c40-43cc-b144-257d6e4385af"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.341428 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljvkg\" (UniqueName: \"kubernetes.io/projected/93845871-6c40-43cc-b144-257d6e4385af-kube-api-access-ljvkg\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.341464 5025 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93845871-6c40-43cc-b144-257d6e4385af-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.341480 5025 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93845871-6c40-43cc-b144-257d6e4385af-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.792842 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" event={"ID":"93845871-6c40-43cc-b144-257d6e4385af","Type":"ContainerDied","Data":"03d08843008f6e0944870556cd80c87c4beb9fa3276db5872483d733e0a2f346"} Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.792893 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03d08843008f6e0944870556cd80c87c4beb9fa3276db5872483d733e0a2f346" Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.792952 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq" Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.797148 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vv75n" event={"ID":"d2f03f7c-2b0c-42f3-b2ec-121f95090044","Type":"ContainerStarted","Data":"49b001690cee6f7183fc5c47edb3a40b4762b665ccb3cb11fd61a3236e447d8d"} Oct 04 11:00:03 crc kubenswrapper[5025]: I1004 11:00:03.821520 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vv75n" podStartSLOduration=2.995244537 podStartE2EDuration="5.821503749s" podCreationTimestamp="2025-10-04 10:59:58 +0000 UTC" firstStartedPulling="2025-10-04 10:59:59.746033244 +0000 UTC m=+1528.171000124" lastFinishedPulling="2025-10-04 11:00:02.572292446 +0000 UTC m=+1530.997259336" observedRunningTime="2025-10-04 11:00:03.814479504 +0000 UTC m=+1532.239446404" watchObservedRunningTime="2025-10-04 11:00:03.821503749 +0000 UTC m=+1532.246470629" Oct 04 11:00:09 crc kubenswrapper[5025]: I1004 11:00:09.164123 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vv75n" Oct 04 11:00:09 crc kubenswrapper[5025]: I1004 11:00:09.164884 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vv75n" Oct 04 11:00:09 crc kubenswrapper[5025]: I1004 11:00:09.222778 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vv75n" Oct 04 11:00:09 crc kubenswrapper[5025]: I1004 11:00:09.936162 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vv75n" Oct 04 11:00:09 crc kubenswrapper[5025]: I1004 11:00:09.988598 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vv75n"] Oct 04 11:00:11 crc kubenswrapper[5025]: I1004 11:00:11.891573 5025 generic.go:334] "Generic (PLEG): container finished" podID="28e59cf5-a704-4eb4-94e7-84a90ceca38d" containerID="fc3ae3b41199f7fd57cf5a1867dcfe59b5935d4a5ccc366b98eb9dffc66bd77c" exitCode=0 Oct 04 11:00:11 crc kubenswrapper[5025]: I1004 11:00:11.891695 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" event={"ID":"28e59cf5-a704-4eb4-94e7-84a90ceca38d","Type":"ContainerDied","Data":"fc3ae3b41199f7fd57cf5a1867dcfe59b5935d4a5ccc366b98eb9dffc66bd77c"} Oct 04 11:00:11 crc kubenswrapper[5025]: I1004 11:00:11.892477 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vv75n" podUID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerName="registry-server" containerID="cri-o://49b001690cee6f7183fc5c47edb3a40b4762b665ccb3cb11fd61a3236e447d8d" gracePeriod=2 Oct 04 11:00:12 crc kubenswrapper[5025]: I1004 11:00:12.905068 5025 generic.go:334] "Generic (PLEG): container finished" podID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerID="49b001690cee6f7183fc5c47edb3a40b4762b665ccb3cb11fd61a3236e447d8d" exitCode=0 Oct 04 11:00:12 crc kubenswrapper[5025]: I1004 11:00:12.905231 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vv75n" event={"ID":"d2f03f7c-2b0c-42f3-b2ec-121f95090044","Type":"ContainerDied","Data":"49b001690cee6f7183fc5c47edb3a40b4762b665ccb3cb11fd61a3236e447d8d"} Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.333456 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.428194 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8jhr\" (UniqueName: \"kubernetes.io/projected/28e59cf5-a704-4eb4-94e7-84a90ceca38d-kube-api-access-m8jhr\") pod \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.428281 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-ssh-key\") pod \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.428315 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-bootstrap-combined-ca-bundle\") pod \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.428487 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-inventory\") pod \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\" (UID: \"28e59cf5-a704-4eb4-94e7-84a90ceca38d\") " Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.440289 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "28e59cf5-a704-4eb4-94e7-84a90ceca38d" (UID: "28e59cf5-a704-4eb4-94e7-84a90ceca38d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.444135 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28e59cf5-a704-4eb4-94e7-84a90ceca38d-kube-api-access-m8jhr" (OuterVolumeSpecName: "kube-api-access-m8jhr") pod "28e59cf5-a704-4eb4-94e7-84a90ceca38d" (UID: "28e59cf5-a704-4eb4-94e7-84a90ceca38d"). InnerVolumeSpecName "kube-api-access-m8jhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.454751 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "28e59cf5-a704-4eb4-94e7-84a90ceca38d" (UID: "28e59cf5-a704-4eb4-94e7-84a90ceca38d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.455147 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-inventory" (OuterVolumeSpecName: "inventory") pod "28e59cf5-a704-4eb4-94e7-84a90ceca38d" (UID: "28e59cf5-a704-4eb4-94e7-84a90ceca38d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.530048 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vv75n" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.530397 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.530429 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8jhr\" (UniqueName: \"kubernetes.io/projected/28e59cf5-a704-4eb4-94e7-84a90ceca38d-kube-api-access-m8jhr\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.530442 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.530453 5025 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28e59cf5-a704-4eb4-94e7-84a90ceca38d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.631673 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-utilities\") pod \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.631750 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-catalog-content\") pod \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.631848 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jp2ls\" (UniqueName: \"kubernetes.io/projected/d2f03f7c-2b0c-42f3-b2ec-121f95090044-kube-api-access-jp2ls\") pod \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\" (UID: \"d2f03f7c-2b0c-42f3-b2ec-121f95090044\") " Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.633457 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-utilities" (OuterVolumeSpecName: "utilities") pod "d2f03f7c-2b0c-42f3-b2ec-121f95090044" (UID: "d2f03f7c-2b0c-42f3-b2ec-121f95090044"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.635121 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2f03f7c-2b0c-42f3-b2ec-121f95090044-kube-api-access-jp2ls" (OuterVolumeSpecName: "kube-api-access-jp2ls") pod "d2f03f7c-2b0c-42f3-b2ec-121f95090044" (UID: "d2f03f7c-2b0c-42f3-b2ec-121f95090044"). InnerVolumeSpecName "kube-api-access-jp2ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.678665 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d2f03f7c-2b0c-42f3-b2ec-121f95090044" (UID: "d2f03f7c-2b0c-42f3-b2ec-121f95090044"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.734471 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.734533 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d2f03f7c-2b0c-42f3-b2ec-121f95090044-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.734558 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jp2ls\" (UniqueName: \"kubernetes.io/projected/d2f03f7c-2b0c-42f3-b2ec-121f95090044-kube-api-access-jp2ls\") on node \"crc\" DevicePath \"\"" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.924563 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vv75n" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.924565 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vv75n" event={"ID":"d2f03f7c-2b0c-42f3-b2ec-121f95090044","Type":"ContainerDied","Data":"a9cbac56a4965f713d7d20e0c3096631e93106fb8aee0d01817f483b3a99edd0"} Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.924644 5025 scope.go:117] "RemoveContainer" containerID="49b001690cee6f7183fc5c47edb3a40b4762b665ccb3cb11fd61a3236e447d8d" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.934377 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" event={"ID":"28e59cf5-a704-4eb4-94e7-84a90ceca38d","Type":"ContainerDied","Data":"3d4de37d49f9797749516cffad9ee251f299d460c677c045135c7535852b72a7"} Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.934449 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d4de37d49f9797749516cffad9ee251f299d460c677c045135c7535852b72a7" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.934549 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.964874 5025 scope.go:117] "RemoveContainer" containerID="e039b4c7959039bf23c332d017663b7f9e2f31c714c062c63b034c891463e525" Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.982733 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vv75n"] Oct 04 11:00:13 crc kubenswrapper[5025]: I1004 11:00:13.996293 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vv75n"] Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.016534 5025 scope.go:117] "RemoveContainer" containerID="55557599490bd10653e5def755171c31b4e41d8b7c82eb997744578ce8272bb9" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.059410 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v"] Oct 04 11:00:14 crc kubenswrapper[5025]: E1004 11:00:14.061539 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerName="registry-server" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.061661 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerName="registry-server" Oct 04 11:00:14 crc kubenswrapper[5025]: E1004 11:00:14.061677 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerName="extract-content" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.061684 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerName="extract-content" Oct 04 11:00:14 crc kubenswrapper[5025]: E1004 11:00:14.061704 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28e59cf5-a704-4eb4-94e7-84a90ceca38d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.061710 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="28e59cf5-a704-4eb4-94e7-84a90ceca38d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 11:00:14 crc kubenswrapper[5025]: E1004 11:00:14.061723 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerName="extract-utilities" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.061729 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerName="extract-utilities" Oct 04 11:00:14 crc kubenswrapper[5025]: E1004 11:00:14.061752 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93845871-6c40-43cc-b144-257d6e4385af" containerName="collect-profiles" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.061757 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="93845871-6c40-43cc-b144-257d6e4385af" containerName="collect-profiles" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.062387 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="28e59cf5-a704-4eb4-94e7-84a90ceca38d" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.062420 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" containerName="registry-server" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.062432 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="93845871-6c40-43cc-b144-257d6e4385af" containerName="collect-profiles" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.063388 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v"] Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.063473 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.065151 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.066911 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.067267 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.067190 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.142519 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p62k\" (UniqueName: \"kubernetes.io/projected/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-kube-api-access-6p62k\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.142624 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.142677 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.244250 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p62k\" (UniqueName: \"kubernetes.io/projected/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-kube-api-access-6p62k\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.244397 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.244485 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.250202 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.250749 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.266190 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p62k\" (UniqueName: \"kubernetes.io/projected/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-kube-api-access-6p62k\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.387148 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.431729 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2f03f7c-2b0c-42f3-b2ec-121f95090044" path="/var/lib/kubelet/pods/d2f03f7c-2b0c-42f3-b2ec-121f95090044/volumes" Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.927570 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v"] Oct 04 11:00:14 crc kubenswrapper[5025]: I1004 11:00:14.946347 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" event={"ID":"aca99a65-247d-49eb-8bb8-4016fe9fdfd0","Type":"ContainerStarted","Data":"d18d228ea80c9d62aec0754c074ec28b0810116ed6c5302b04a55fc0789c9dfc"} Oct 04 11:00:15 crc kubenswrapper[5025]: I1004 11:00:15.959566 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" event={"ID":"aca99a65-247d-49eb-8bb8-4016fe9fdfd0","Type":"ContainerStarted","Data":"cdaa32c780a1bdf5a71df19050fa7c0caf38a4ed7bd32939437e96913488a357"} Oct 04 11:00:15 crc kubenswrapper[5025]: I1004 11:00:15.982299 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" podStartSLOduration=2.441472729 podStartE2EDuration="2.982279424s" podCreationTimestamp="2025-10-04 11:00:13 +0000 UTC" firstStartedPulling="2025-10-04 11:00:14.932474663 +0000 UTC m=+1543.357441583" lastFinishedPulling="2025-10-04 11:00:15.473281388 +0000 UTC m=+1543.898248278" observedRunningTime="2025-10-04 11:00:15.978935746 +0000 UTC m=+1544.403902636" watchObservedRunningTime="2025-10-04 11:00:15.982279424 +0000 UTC m=+1544.407246304" Oct 04 11:00:47 crc kubenswrapper[5025]: I1004 11:00:47.380068 5025 scope.go:117] "RemoveContainer" containerID="14743e2481323cb7ef3ffe0099e6662bca90c3c14cca7c7625b0dbd06bbaa8f6" Oct 04 11:00:47 crc kubenswrapper[5025]: I1004 11:00:47.404580 5025 scope.go:117] "RemoveContainer" containerID="08c5ac95c8a494b2c84c8f29df3f940bd79ee23fee759ecba9b6c7c8ac4066d7" Oct 04 11:00:47 crc kubenswrapper[5025]: I1004 11:00:47.425642 5025 scope.go:117] "RemoveContainer" containerID="6c76fe1b0b99182de30bf2bae872be46b21b7795e4170461121f739587d9b833" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.161892 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29326261-55b5l"] Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.164459 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.173745 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29326261-55b5l"] Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.347975 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-config-data\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.348361 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rzkn\" (UniqueName: \"kubernetes.io/projected/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-kube-api-access-9rzkn\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.348394 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-fernet-keys\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.348436 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-combined-ca-bundle\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.449648 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-config-data\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.449707 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rzkn\" (UniqueName: \"kubernetes.io/projected/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-kube-api-access-9rzkn\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.449726 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-fernet-keys\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.449758 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-combined-ca-bundle\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.458157 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-fernet-keys\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.459555 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-combined-ca-bundle\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.459748 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-config-data\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.480947 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rzkn\" (UniqueName: \"kubernetes.io/projected/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-kube-api-access-9rzkn\") pod \"keystone-cron-29326261-55b5l\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.499099 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:00 crc kubenswrapper[5025]: W1004 11:01:00.977899 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc05eaa2_72fa_4aef_8a86_a1bae9d427c4.slice/crio-79a053216da02304a496e289c8745e46d1b3813a55475ffb3a5828c758e5be2e WatchSource:0}: Error finding container 79a053216da02304a496e289c8745e46d1b3813a55475ffb3a5828c758e5be2e: Status 404 returned error can't find the container with id 79a053216da02304a496e289c8745e46d1b3813a55475ffb3a5828c758e5be2e Oct 04 11:01:00 crc kubenswrapper[5025]: I1004 11:01:00.981587 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29326261-55b5l"] Oct 04 11:01:01 crc kubenswrapper[5025]: I1004 11:01:01.509481 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326261-55b5l" event={"ID":"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4","Type":"ContainerStarted","Data":"79a053216da02304a496e289c8745e46d1b3813a55475ffb3a5828c758e5be2e"} Oct 04 11:01:02 crc kubenswrapper[5025]: I1004 11:01:02.517978 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326261-55b5l" event={"ID":"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4","Type":"ContainerStarted","Data":"889b58a0121d29e022d66f1f903c3f36df9631d0b68c14ec186e4e0cc26d5c7b"} Oct 04 11:01:02 crc kubenswrapper[5025]: I1004 11:01:02.538704 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29326261-55b5l" podStartSLOduration=2.5386868270000003 podStartE2EDuration="2.538686827s" podCreationTimestamp="2025-10-04 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:01:02.532052823 +0000 UTC m=+1590.957019713" watchObservedRunningTime="2025-10-04 11:01:02.538686827 +0000 UTC m=+1590.963653707" Oct 04 11:01:03 crc kubenswrapper[5025]: I1004 11:01:03.532334 5025 generic.go:334] "Generic (PLEG): container finished" podID="fc05eaa2-72fa-4aef-8a86-a1bae9d427c4" containerID="889b58a0121d29e022d66f1f903c3f36df9631d0b68c14ec186e4e0cc26d5c7b" exitCode=0 Oct 04 11:01:03 crc kubenswrapper[5025]: I1004 11:01:03.532710 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326261-55b5l" event={"ID":"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4","Type":"ContainerDied","Data":"889b58a0121d29e022d66f1f903c3f36df9631d0b68c14ec186e4e0cc26d5c7b"} Oct 04 11:01:04 crc kubenswrapper[5025]: I1004 11:01:04.950909 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.063511 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rzkn\" (UniqueName: \"kubernetes.io/projected/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-kube-api-access-9rzkn\") pod \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.064097 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-fernet-keys\") pod \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.064347 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-config-data\") pod \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.064641 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-combined-ca-bundle\") pod \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\" (UID: \"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4\") " Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.069276 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fc05eaa2-72fa-4aef-8a86-a1bae9d427c4" (UID: "fc05eaa2-72fa-4aef-8a86-a1bae9d427c4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.069963 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-kube-api-access-9rzkn" (OuterVolumeSpecName: "kube-api-access-9rzkn") pod "fc05eaa2-72fa-4aef-8a86-a1bae9d427c4" (UID: "fc05eaa2-72fa-4aef-8a86-a1bae9d427c4"). InnerVolumeSpecName "kube-api-access-9rzkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.090413 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc05eaa2-72fa-4aef-8a86-a1bae9d427c4" (UID: "fc05eaa2-72fa-4aef-8a86-a1bae9d427c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.132886 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-config-data" (OuterVolumeSpecName: "config-data") pod "fc05eaa2-72fa-4aef-8a86-a1bae9d427c4" (UID: "fc05eaa2-72fa-4aef-8a86-a1bae9d427c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.167609 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.167646 5025 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.167662 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rzkn\" (UniqueName: \"kubernetes.io/projected/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-kube-api-access-9rzkn\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.167674 5025 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fc05eaa2-72fa-4aef-8a86-a1bae9d427c4-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.557216 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29326261-55b5l" event={"ID":"fc05eaa2-72fa-4aef-8a86-a1bae9d427c4","Type":"ContainerDied","Data":"79a053216da02304a496e289c8745e46d1b3813a55475ffb3a5828c758e5be2e"} Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.557259 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79a053216da02304a496e289c8745e46d1b3813a55475ffb3a5828c758e5be2e" Oct 04 11:01:05 crc kubenswrapper[5025]: I1004 11:01:05.557280 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29326261-55b5l" Oct 04 11:01:06 crc kubenswrapper[5025]: I1004 11:01:06.051722 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-fzr5t"] Oct 04 11:01:06 crc kubenswrapper[5025]: I1004 11:01:06.064215 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-fzr5t"] Oct 04 11:01:06 crc kubenswrapper[5025]: I1004 11:01:06.434036 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39ee3358-4ace-4a12-93c8-90f0fbead7c5" path="/var/lib/kubelet/pods/39ee3358-4ace-4a12-93c8-90f0fbead7c5/volumes" Oct 04 11:01:11 crc kubenswrapper[5025]: I1004 11:01:11.025356 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-k4fff"] Oct 04 11:01:11 crc kubenswrapper[5025]: I1004 11:01:11.036465 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-jhjgk"] Oct 04 11:01:11 crc kubenswrapper[5025]: I1004 11:01:11.046011 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-jhjgk"] Oct 04 11:01:11 crc kubenswrapper[5025]: I1004 11:01:11.055566 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-k4fff"] Oct 04 11:01:12 crc kubenswrapper[5025]: I1004 11:01:12.428914 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c58f978-1c60-4b77-b6cc-519d0b28f447" path="/var/lib/kubelet/pods/8c58f978-1c60-4b77-b6cc-519d0b28f447/volumes" Oct 04 11:01:12 crc kubenswrapper[5025]: I1004 11:01:12.430117 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac1e5058-70ea-478e-b265-d737830e6daa" path="/var/lib/kubelet/pods/ac1e5058-70ea-478e-b265-d737830e6daa/volumes" Oct 04 11:01:14 crc kubenswrapper[5025]: I1004 11:01:14.713762 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:01:14 crc kubenswrapper[5025]: I1004 11:01:14.714174 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:01:16 crc kubenswrapper[5025]: I1004 11:01:16.024392 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-e3d9-account-create-hrch9"] Oct 04 11:01:16 crc kubenswrapper[5025]: I1004 11:01:16.031584 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-e3d9-account-create-hrch9"] Oct 04 11:01:16 crc kubenswrapper[5025]: I1004 11:01:16.428826 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06c89e30-b6f2-4416-a976-e0661bd1fa07" path="/var/lib/kubelet/pods/06c89e30-b6f2-4416-a976-e0661bd1fa07/volumes" Oct 04 11:01:20 crc kubenswrapper[5025]: I1004 11:01:20.036055 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-ccc4-account-create-pl9j2"] Oct 04 11:01:20 crc kubenswrapper[5025]: I1004 11:01:20.050897 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-ccc4-account-create-pl9j2"] Oct 04 11:01:20 crc kubenswrapper[5025]: I1004 11:01:20.424312 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dad9687f-030b-4379-ba71-965b512fadaa" path="/var/lib/kubelet/pods/dad9687f-030b-4379-ba71-965b512fadaa/volumes" Oct 04 11:01:21 crc kubenswrapper[5025]: I1004 11:01:21.040742 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-9b7d-account-create-q2442"] Oct 04 11:01:21 crc kubenswrapper[5025]: I1004 11:01:21.051266 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-9b7d-account-create-q2442"] Oct 04 11:01:22 crc kubenswrapper[5025]: I1004 11:01:22.430945 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="083601ee-774b-4e6d-a8e0-debebeb9011a" path="/var/lib/kubelet/pods/083601ee-774b-4e6d-a8e0-debebeb9011a/volumes" Oct 04 11:01:33 crc kubenswrapper[5025]: I1004 11:01:33.039662 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-22h8p"] Oct 04 11:01:33 crc kubenswrapper[5025]: I1004 11:01:33.056195 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-bffxf"] Oct 04 11:01:33 crc kubenswrapper[5025]: I1004 11:01:33.065978 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-22h8p"] Oct 04 11:01:33 crc kubenswrapper[5025]: I1004 11:01:33.074850 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-bffxf"] Oct 04 11:01:34 crc kubenswrapper[5025]: I1004 11:01:34.428459 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26b3df34-d44c-4f4a-888d-73413cd3a358" path="/var/lib/kubelet/pods/26b3df34-d44c-4f4a-888d-73413cd3a358/volumes" Oct 04 11:01:34 crc kubenswrapper[5025]: I1004 11:01:34.429543 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05" path="/var/lib/kubelet/pods/31c67d5a-eb1f-4fc0-b3bd-1bc89799cb05/volumes" Oct 04 11:01:37 crc kubenswrapper[5025]: I1004 11:01:37.039725 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-xlm27"] Oct 04 11:01:37 crc kubenswrapper[5025]: I1004 11:01:37.046660 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-xlm27"] Oct 04 11:01:38 crc kubenswrapper[5025]: I1004 11:01:38.431311 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98f93315-6097-4365-9972-03e9559f8c32" path="/var/lib/kubelet/pods/98f93315-6097-4365-9972-03e9559f8c32/volumes" Oct 04 11:01:44 crc kubenswrapper[5025]: I1004 11:01:44.713942 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:01:44 crc kubenswrapper[5025]: I1004 11:01:44.714650 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.036548 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-frhrp"] Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.050951 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c3e5-account-create-sxqrv"] Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.059768 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-1b98-account-create-c872x"] Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.070562 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c3e5-account-create-sxqrv"] Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.080285 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-frhrp"] Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.087635 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-1b98-account-create-c872x"] Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.518371 5025 scope.go:117] "RemoveContainer" containerID="19bbbec6b4c169ccdae4e47b4e0dcbfaf353e25657dce3380401556152a324fc" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.553098 5025 scope.go:117] "RemoveContainer" containerID="aa7acc7ce1d3f4d7db14c73df69839aa8cac991c3589c934777f00ef5e75e4bf" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.618562 5025 scope.go:117] "RemoveContainer" containerID="e9eaba596fa46d38b6f26e181da1df105847685501383c957151836c04ef4d5d" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.678239 5025 scope.go:117] "RemoveContainer" containerID="736c3219df1ef85213d8e33524f7aa40ff317e2c6d95580b6fd88acb6617cecb" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.707494 5025 scope.go:117] "RemoveContainer" containerID="4d306b6d67206c37b584322b76a665a377d05d028e7cb1f5b991b7c1836e5239" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.735460 5025 scope.go:117] "RemoveContainer" containerID="df0305ddf1cd74da92eed6ad7d1953ed912edde49e3d7488ae3ed3e9baa705c7" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.775523 5025 scope.go:117] "RemoveContainer" containerID="df5cd369869c48699d0521f44fb93465ad2f421b591b1e128112798d7d497569" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.833587 5025 scope.go:117] "RemoveContainer" containerID="f4b369e9a8aabe353e58f5866e604c4783c6390252fa27d24258c2e9b3b60e43" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.852120 5025 scope.go:117] "RemoveContainer" containerID="29e9d50ccd3e83d63217623c97ff5a6fc362c1de005e9a19d50f8373368ab1b3" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.878200 5025 scope.go:117] "RemoveContainer" containerID="d66c64aa24974228191f78f1791b62a463234c3fde5fa924e3f8d1a845a65324" Oct 04 11:01:47 crc kubenswrapper[5025]: I1004 11:01:47.899694 5025 scope.go:117] "RemoveContainer" containerID="3b94de62d89966ce710fcb794f59a3425d495d60754a1e0b4fcced534c0223d4" Oct 04 11:01:48 crc kubenswrapper[5025]: I1004 11:01:48.039341 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-ctpv7"] Oct 04 11:01:48 crc kubenswrapper[5025]: I1004 11:01:48.049482 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-ctpv7"] Oct 04 11:01:48 crc kubenswrapper[5025]: I1004 11:01:48.431258 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1270a45a-df30-4138-8270-78174433f9cb" path="/var/lib/kubelet/pods/1270a45a-df30-4138-8270-78174433f9cb/volumes" Oct 04 11:01:48 crc kubenswrapper[5025]: I1004 11:01:48.432405 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="298276d3-1df7-4934-88b3-cfcbfbc0fd51" path="/var/lib/kubelet/pods/298276d3-1df7-4934-88b3-cfcbfbc0fd51/volumes" Oct 04 11:01:48 crc kubenswrapper[5025]: I1004 11:01:48.433600 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4323bbf4-8801-411a-8fce-bd4a077180da" path="/var/lib/kubelet/pods/4323bbf4-8801-411a-8fce-bd4a077180da/volumes" Oct 04 11:01:48 crc kubenswrapper[5025]: I1004 11:01:48.435857 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d3aea52-35ad-49e4-93fb-465cd111f845" path="/var/lib/kubelet/pods/4d3aea52-35ad-49e4-93fb-465cd111f845/volumes" Oct 04 11:01:49 crc kubenswrapper[5025]: I1004 11:01:49.037693 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c114-account-create-ttrlz"] Oct 04 11:01:49 crc kubenswrapper[5025]: I1004 11:01:49.049293 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c114-account-create-ttrlz"] Oct 04 11:01:50 crc kubenswrapper[5025]: I1004 11:01:50.429944 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce76a45b-1519-4a6e-9d7a-7a33bebadc13" path="/var/lib/kubelet/pods/ce76a45b-1519-4a6e-9d7a-7a33bebadc13/volumes" Oct 04 11:01:53 crc kubenswrapper[5025]: I1004 11:01:53.100235 5025 generic.go:334] "Generic (PLEG): container finished" podID="aca99a65-247d-49eb-8bb8-4016fe9fdfd0" containerID="cdaa32c780a1bdf5a71df19050fa7c0caf38a4ed7bd32939437e96913488a357" exitCode=0 Oct 04 11:01:53 crc kubenswrapper[5025]: I1004 11:01:53.100303 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" event={"ID":"aca99a65-247d-49eb-8bb8-4016fe9fdfd0","Type":"ContainerDied","Data":"cdaa32c780a1bdf5a71df19050fa7c0caf38a4ed7bd32939437e96913488a357"} Oct 04 11:01:54 crc kubenswrapper[5025]: I1004 11:01:54.667159 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:01:54 crc kubenswrapper[5025]: I1004 11:01:54.860945 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p62k\" (UniqueName: \"kubernetes.io/projected/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-kube-api-access-6p62k\") pod \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " Oct 04 11:01:54 crc kubenswrapper[5025]: I1004 11:01:54.861079 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-ssh-key\") pod \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " Oct 04 11:01:54 crc kubenswrapper[5025]: I1004 11:01:54.861320 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-inventory\") pod \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\" (UID: \"aca99a65-247d-49eb-8bb8-4016fe9fdfd0\") " Oct 04 11:01:54 crc kubenswrapper[5025]: I1004 11:01:54.867333 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-kube-api-access-6p62k" (OuterVolumeSpecName: "kube-api-access-6p62k") pod "aca99a65-247d-49eb-8bb8-4016fe9fdfd0" (UID: "aca99a65-247d-49eb-8bb8-4016fe9fdfd0"). InnerVolumeSpecName "kube-api-access-6p62k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:01:54 crc kubenswrapper[5025]: I1004 11:01:54.898836 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-inventory" (OuterVolumeSpecName: "inventory") pod "aca99a65-247d-49eb-8bb8-4016fe9fdfd0" (UID: "aca99a65-247d-49eb-8bb8-4016fe9fdfd0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:01:54 crc kubenswrapper[5025]: I1004 11:01:54.906040 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "aca99a65-247d-49eb-8bb8-4016fe9fdfd0" (UID: "aca99a65-247d-49eb-8bb8-4016fe9fdfd0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:01:54 crc kubenswrapper[5025]: I1004 11:01:54.965139 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:54 crc kubenswrapper[5025]: I1004 11:01:54.965199 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p62k\" (UniqueName: \"kubernetes.io/projected/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-kube-api-access-6p62k\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:54 crc kubenswrapper[5025]: I1004 11:01:54.965221 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/aca99a65-247d-49eb-8bb8-4016fe9fdfd0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.125108 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" event={"ID":"aca99a65-247d-49eb-8bb8-4016fe9fdfd0","Type":"ContainerDied","Data":"d18d228ea80c9d62aec0754c074ec28b0810116ed6c5302b04a55fc0789c9dfc"} Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.125156 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d18d228ea80c9d62aec0754c074ec28b0810116ed6c5302b04a55fc0789c9dfc" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.125205 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.225295 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw"] Oct 04 11:01:55 crc kubenswrapper[5025]: E1004 11:01:55.225867 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aca99a65-247d-49eb-8bb8-4016fe9fdfd0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.225893 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="aca99a65-247d-49eb-8bb8-4016fe9fdfd0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 11:01:55 crc kubenswrapper[5025]: E1004 11:01:55.225906 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc05eaa2-72fa-4aef-8a86-a1bae9d427c4" containerName="keystone-cron" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.225914 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc05eaa2-72fa-4aef-8a86-a1bae9d427c4" containerName="keystone-cron" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.226391 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc05eaa2-72fa-4aef-8a86-a1bae9d427c4" containerName="keystone-cron" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.226409 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="aca99a65-247d-49eb-8bb8-4016fe9fdfd0" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.238962 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.242602 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw"] Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.243715 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.243856 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.243956 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.244274 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.372904 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.373091 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4bxw\" (UniqueName: \"kubernetes.io/projected/bfac76b8-067d-4dc5-8d95-930cf9f08863-kube-api-access-x4bxw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.373128 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.475486 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4bxw\" (UniqueName: \"kubernetes.io/projected/bfac76b8-067d-4dc5-8d95-930cf9f08863-kube-api-access-x4bxw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.475581 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.475666 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.481401 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.482120 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.496068 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4bxw\" (UniqueName: \"kubernetes.io/projected/bfac76b8-067d-4dc5-8d95-930cf9f08863-kube-api-access-x4bxw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.560791 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.954406 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw"] Oct 04 11:01:55 crc kubenswrapper[5025]: I1004 11:01:55.965734 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:01:56 crc kubenswrapper[5025]: I1004 11:01:56.134698 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" event={"ID":"bfac76b8-067d-4dc5-8d95-930cf9f08863","Type":"ContainerStarted","Data":"76cf248a5d4a76cb38afdc2784afbe42550fec3f33cfd9bffe2a1a0a8ca46a59"} Oct 04 11:01:57 crc kubenswrapper[5025]: I1004 11:01:57.143996 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" event={"ID":"bfac76b8-067d-4dc5-8d95-930cf9f08863","Type":"ContainerStarted","Data":"420dd815f0b4f50a4fd21a6d729ae0ed38e0e04f19a0aa3b55688faa346545ef"} Oct 04 11:02:14 crc kubenswrapper[5025]: I1004 11:02:14.713983 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:02:14 crc kubenswrapper[5025]: I1004 11:02:14.714690 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:02:14 crc kubenswrapper[5025]: I1004 11:02:14.714754 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 11:02:14 crc kubenswrapper[5025]: I1004 11:02:14.715786 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:02:14 crc kubenswrapper[5025]: I1004 11:02:14.715928 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" gracePeriod=600 Oct 04 11:02:14 crc kubenswrapper[5025]: E1004 11:02:14.847742 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:02:15 crc kubenswrapper[5025]: I1004 11:02:15.349926 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" exitCode=0 Oct 04 11:02:15 crc kubenswrapper[5025]: I1004 11:02:15.349994 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634"} Oct 04 11:02:15 crc kubenswrapper[5025]: I1004 11:02:15.350508 5025 scope.go:117] "RemoveContainer" containerID="56f39d1552d0ae55dd9141a0ac88b016c83fd194f8df926b4fe6a98c6db80007" Oct 04 11:02:15 crc kubenswrapper[5025]: I1004 11:02:15.352795 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:02:15 crc kubenswrapper[5025]: E1004 11:02:15.353840 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:02:15 crc kubenswrapper[5025]: I1004 11:02:15.378010 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" podStartSLOduration=19.958020379 podStartE2EDuration="20.377983589s" podCreationTimestamp="2025-10-04 11:01:55 +0000 UTC" firstStartedPulling="2025-10-04 11:01:55.965478682 +0000 UTC m=+1644.390445562" lastFinishedPulling="2025-10-04 11:01:56.385441852 +0000 UTC m=+1644.810408772" observedRunningTime="2025-10-04 11:01:57.172131737 +0000 UTC m=+1645.597098617" watchObservedRunningTime="2025-10-04 11:02:15.377983589 +0000 UTC m=+1663.802950509" Oct 04 11:02:18 crc kubenswrapper[5025]: I1004 11:02:18.038728 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-25rkf"] Oct 04 11:02:18 crc kubenswrapper[5025]: I1004 11:02:18.046875 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-25rkf"] Oct 04 11:02:18 crc kubenswrapper[5025]: I1004 11:02:18.425681 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e307d63-f22a-4e97-9385-773e9ce7f4fd" path="/var/lib/kubelet/pods/8e307d63-f22a-4e97-9385-773e9ce7f4fd/volumes" Oct 04 11:02:19 crc kubenswrapper[5025]: I1004 11:02:19.034213 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-g94ln"] Oct 04 11:02:19 crc kubenswrapper[5025]: I1004 11:02:19.041663 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-g94ln"] Oct 04 11:02:20 crc kubenswrapper[5025]: I1004 11:02:20.428918 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="565cc52d-cdcf-4d59-82b5-3d71804b727a" path="/var/lib/kubelet/pods/565cc52d-cdcf-4d59-82b5-3d71804b727a/volumes" Oct 04 11:02:26 crc kubenswrapper[5025]: I1004 11:02:26.049453 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dncjm"] Oct 04 11:02:26 crc kubenswrapper[5025]: I1004 11:02:26.056116 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dncjm"] Oct 04 11:02:26 crc kubenswrapper[5025]: I1004 11:02:26.422840 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a4861ec-2e6e-408d-9b42-5ce3248fd640" path="/var/lib/kubelet/pods/4a4861ec-2e6e-408d-9b42-5ce3248fd640/volumes" Oct 04 11:02:28 crc kubenswrapper[5025]: I1004 11:02:28.030817 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-sfm5z"] Oct 04 11:02:28 crc kubenswrapper[5025]: I1004 11:02:28.039070 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-sfm5z"] Oct 04 11:02:28 crc kubenswrapper[5025]: I1004 11:02:28.412186 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:02:28 crc kubenswrapper[5025]: E1004 11:02:28.412995 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:02:28 crc kubenswrapper[5025]: I1004 11:02:28.424751 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5d127d5-e426-48fe-bb22-c811de5c801d" path="/var/lib/kubelet/pods/b5d127d5-e426-48fe-bb22-c811de5c801d/volumes" Oct 04 11:02:38 crc kubenswrapper[5025]: I1004 11:02:38.036213 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-xtspb"] Oct 04 11:02:38 crc kubenswrapper[5025]: I1004 11:02:38.049699 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-xtspb"] Oct 04 11:02:38 crc kubenswrapper[5025]: I1004 11:02:38.421479 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="694e381e-49b8-49f1-a5ae-7aa3b72d8fee" path="/var/lib/kubelet/pods/694e381e-49b8-49f1-a5ae-7aa3b72d8fee/volumes" Oct 04 11:02:43 crc kubenswrapper[5025]: I1004 11:02:43.410795 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:02:43 crc kubenswrapper[5025]: E1004 11:02:43.412370 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:02:48 crc kubenswrapper[5025]: I1004 11:02:48.103510 5025 scope.go:117] "RemoveContainer" containerID="2b783084ad2ac2c644297be5333b2f46bdda1ea8bdaa25217af12d8ae5abc93a" Oct 04 11:02:48 crc kubenswrapper[5025]: I1004 11:02:48.169639 5025 scope.go:117] "RemoveContainer" containerID="5df93cf41319e8d922adf4b51dfb0ad272c02b04679fff4cb8b2158749bf220e" Oct 04 11:02:48 crc kubenswrapper[5025]: I1004 11:02:48.204512 5025 scope.go:117] "RemoveContainer" containerID="149b5de7d7e8cea0b473cf87d614843df1451b44ccf2ee3c07c16f2c2af9005e" Oct 04 11:02:48 crc kubenswrapper[5025]: I1004 11:02:48.251950 5025 scope.go:117] "RemoveContainer" containerID="22c2e8b34fd6e772d90658dccb31c3f9b4484035315f76d6b01e48d3d56d0157" Oct 04 11:02:48 crc kubenswrapper[5025]: I1004 11:02:48.303551 5025 scope.go:117] "RemoveContainer" containerID="27f95935eab866d09895b9f363cc4d687040656a7e9cf672deb677ef1396860a" Oct 04 11:02:48 crc kubenswrapper[5025]: I1004 11:02:48.366244 5025 scope.go:117] "RemoveContainer" containerID="6bb3b61562fd3fec493b942be4e930f472a30c9bb280ebef518fa016041ed7b7" Oct 04 11:02:48 crc kubenswrapper[5025]: I1004 11:02:48.409640 5025 scope.go:117] "RemoveContainer" containerID="c827830940a0fdc33ac2dd558980c09f9def779177524946034990802b8b36a5" Oct 04 11:02:48 crc kubenswrapper[5025]: I1004 11:02:48.430783 5025 scope.go:117] "RemoveContainer" containerID="5f0c2f743e61fbddf48cd901dcc0201092cbfbb21add1655d84f84894f70a55e" Oct 04 11:02:48 crc kubenswrapper[5025]: I1004 11:02:48.447413 5025 scope.go:117] "RemoveContainer" containerID="52df45af5dc37cc05f6bc5aec470bfdecd9f2f4c802694476feefddd6e716cf1" Oct 04 11:02:48 crc kubenswrapper[5025]: I1004 11:02:48.477604 5025 scope.go:117] "RemoveContainer" containerID="013fcc42e6f1203c89c3b5e9723bc4b707a00316525e047d9ef581503f1ca4c3" Oct 04 11:02:55 crc kubenswrapper[5025]: I1004 11:02:55.412157 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:02:55 crc kubenswrapper[5025]: E1004 11:02:55.413091 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:03:09 crc kubenswrapper[5025]: I1004 11:03:09.411708 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:03:09 crc kubenswrapper[5025]: E1004 11:03:09.412775 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:03:10 crc kubenswrapper[5025]: I1004 11:03:10.946781 5025 generic.go:334] "Generic (PLEG): container finished" podID="bfac76b8-067d-4dc5-8d95-930cf9f08863" containerID="420dd815f0b4f50a4fd21a6d729ae0ed38e0e04f19a0aa3b55688faa346545ef" exitCode=0 Oct 04 11:03:10 crc kubenswrapper[5025]: I1004 11:03:10.946826 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" event={"ID":"bfac76b8-067d-4dc5-8d95-930cf9f08863","Type":"ContainerDied","Data":"420dd815f0b4f50a4fd21a6d729ae0ed38e0e04f19a0aa3b55688faa346545ef"} Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.389558 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.520146 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4bxw\" (UniqueName: \"kubernetes.io/projected/bfac76b8-067d-4dc5-8d95-930cf9f08863-kube-api-access-x4bxw\") pod \"bfac76b8-067d-4dc5-8d95-930cf9f08863\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.520209 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-inventory\") pod \"bfac76b8-067d-4dc5-8d95-930cf9f08863\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.520415 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-ssh-key\") pod \"bfac76b8-067d-4dc5-8d95-930cf9f08863\" (UID: \"bfac76b8-067d-4dc5-8d95-930cf9f08863\") " Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.526560 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfac76b8-067d-4dc5-8d95-930cf9f08863-kube-api-access-x4bxw" (OuterVolumeSpecName: "kube-api-access-x4bxw") pod "bfac76b8-067d-4dc5-8d95-930cf9f08863" (UID: "bfac76b8-067d-4dc5-8d95-930cf9f08863"). InnerVolumeSpecName "kube-api-access-x4bxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.561536 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-inventory" (OuterVolumeSpecName: "inventory") pod "bfac76b8-067d-4dc5-8d95-930cf9f08863" (UID: "bfac76b8-067d-4dc5-8d95-930cf9f08863"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.573654 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bfac76b8-067d-4dc5-8d95-930cf9f08863" (UID: "bfac76b8-067d-4dc5-8d95-930cf9f08863"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.623554 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.623860 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4bxw\" (UniqueName: \"kubernetes.io/projected/bfac76b8-067d-4dc5-8d95-930cf9f08863-kube-api-access-x4bxw\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.624007 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bfac76b8-067d-4dc5-8d95-930cf9f08863-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.966357 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" event={"ID":"bfac76b8-067d-4dc5-8d95-930cf9f08863","Type":"ContainerDied","Data":"76cf248a5d4a76cb38afdc2784afbe42550fec3f33cfd9bffe2a1a0a8ca46a59"} Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.966417 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76cf248a5d4a76cb38afdc2784afbe42550fec3f33cfd9bffe2a1a0a8ca46a59" Oct 04 11:03:12 crc kubenswrapper[5025]: I1004 11:03:12.967111 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.095098 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk"] Oct 04 11:03:13 crc kubenswrapper[5025]: E1004 11:03:13.095866 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfac76b8-067d-4dc5-8d95-930cf9f08863" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.095909 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfac76b8-067d-4dc5-8d95-930cf9f08863" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.096510 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfac76b8-067d-4dc5-8d95-930cf9f08863" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.097617 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.100493 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.101353 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.101550 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.103171 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.104584 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk"] Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.236470 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.236977 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.237046 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c467m\" (UniqueName: \"kubernetes.io/projected/9f4db205-f2ac-428f-a424-a238cfa9d0be-kube-api-access-c467m\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.338960 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.339064 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c467m\" (UniqueName: \"kubernetes.io/projected/9f4db205-f2ac-428f-a424-a238cfa9d0be-kube-api-access-c467m\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.339097 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.342803 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.345765 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.358831 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c467m\" (UniqueName: \"kubernetes.io/projected/9f4db205-f2ac-428f-a424-a238cfa9d0be-kube-api-access-c467m\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.439860 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:13 crc kubenswrapper[5025]: I1004 11:03:13.986914 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk"] Oct 04 11:03:14 crc kubenswrapper[5025]: I1004 11:03:14.988072 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" event={"ID":"9f4db205-f2ac-428f-a424-a238cfa9d0be","Type":"ContainerStarted","Data":"89961ce017b03c5c52064f6a06cbfa929b97f0c971018edb42947ea32f030a41"} Oct 04 11:03:14 crc kubenswrapper[5025]: I1004 11:03:14.988362 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" event={"ID":"9f4db205-f2ac-428f-a424-a238cfa9d0be","Type":"ContainerStarted","Data":"459efa1a8294269576ce5ed0041aeaa7c0912f8be80d1eed93b1169b986a9a07"} Oct 04 11:03:15 crc kubenswrapper[5025]: I1004 11:03:15.017255 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" podStartSLOduration=1.365316067 podStartE2EDuration="2.017223747s" podCreationTimestamp="2025-10-04 11:03:13 +0000 UTC" firstStartedPulling="2025-10-04 11:03:13.999291962 +0000 UTC m=+1722.424258842" lastFinishedPulling="2025-10-04 11:03:14.651199612 +0000 UTC m=+1723.076166522" observedRunningTime="2025-10-04 11:03:15.006197215 +0000 UTC m=+1723.431164095" watchObservedRunningTime="2025-10-04 11:03:15.017223747 +0000 UTC m=+1723.442190667" Oct 04 11:03:20 crc kubenswrapper[5025]: I1004 11:03:20.411676 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:03:20 crc kubenswrapper[5025]: E1004 11:03:20.413546 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:03:21 crc kubenswrapper[5025]: I1004 11:03:21.059859 5025 generic.go:334] "Generic (PLEG): container finished" podID="9f4db205-f2ac-428f-a424-a238cfa9d0be" containerID="89961ce017b03c5c52064f6a06cbfa929b97f0c971018edb42947ea32f030a41" exitCode=0 Oct 04 11:03:21 crc kubenswrapper[5025]: I1004 11:03:21.059935 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" event={"ID":"9f4db205-f2ac-428f-a424-a238cfa9d0be","Type":"ContainerDied","Data":"89961ce017b03c5c52064f6a06cbfa929b97f0c971018edb42947ea32f030a41"} Oct 04 11:03:22 crc kubenswrapper[5025]: I1004 11:03:22.479343 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:22 crc kubenswrapper[5025]: I1004 11:03:22.656551 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-ssh-key\") pod \"9f4db205-f2ac-428f-a424-a238cfa9d0be\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " Oct 04 11:03:22 crc kubenswrapper[5025]: I1004 11:03:22.656596 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c467m\" (UniqueName: \"kubernetes.io/projected/9f4db205-f2ac-428f-a424-a238cfa9d0be-kube-api-access-c467m\") pod \"9f4db205-f2ac-428f-a424-a238cfa9d0be\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " Oct 04 11:03:22 crc kubenswrapper[5025]: I1004 11:03:22.656774 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-inventory\") pod \"9f4db205-f2ac-428f-a424-a238cfa9d0be\" (UID: \"9f4db205-f2ac-428f-a424-a238cfa9d0be\") " Oct 04 11:03:22 crc kubenswrapper[5025]: I1004 11:03:22.669254 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f4db205-f2ac-428f-a424-a238cfa9d0be-kube-api-access-c467m" (OuterVolumeSpecName: "kube-api-access-c467m") pod "9f4db205-f2ac-428f-a424-a238cfa9d0be" (UID: "9f4db205-f2ac-428f-a424-a238cfa9d0be"). InnerVolumeSpecName "kube-api-access-c467m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:03:22 crc kubenswrapper[5025]: I1004 11:03:22.685820 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-inventory" (OuterVolumeSpecName: "inventory") pod "9f4db205-f2ac-428f-a424-a238cfa9d0be" (UID: "9f4db205-f2ac-428f-a424-a238cfa9d0be"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:03:22 crc kubenswrapper[5025]: I1004 11:03:22.687452 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9f4db205-f2ac-428f-a424-a238cfa9d0be" (UID: "9f4db205-f2ac-428f-a424-a238cfa9d0be"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:03:22 crc kubenswrapper[5025]: I1004 11:03:22.758697 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:22 crc kubenswrapper[5025]: I1004 11:03:22.758724 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f4db205-f2ac-428f-a424-a238cfa9d0be-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:22 crc kubenswrapper[5025]: I1004 11:03:22.758733 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c467m\" (UniqueName: \"kubernetes.io/projected/9f4db205-f2ac-428f-a424-a238cfa9d0be-kube-api-access-c467m\") on node \"crc\" DevicePath \"\"" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.079500 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" event={"ID":"9f4db205-f2ac-428f-a424-a238cfa9d0be","Type":"ContainerDied","Data":"459efa1a8294269576ce5ed0041aeaa7c0912f8be80d1eed93b1169b986a9a07"} Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.079537 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="459efa1a8294269576ce5ed0041aeaa7c0912f8be80d1eed93b1169b986a9a07" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.079552 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.157768 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s"] Oct 04 11:03:23 crc kubenswrapper[5025]: E1004 11:03:23.158254 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f4db205-f2ac-428f-a424-a238cfa9d0be" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.158276 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f4db205-f2ac-428f-a424-a238cfa9d0be" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.158787 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f4db205-f2ac-428f-a424-a238cfa9d0be" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.159584 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.161645 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.166524 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.166703 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.166765 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.168300 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s"] Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.266651 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kjz6s\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.266970 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x28cx\" (UniqueName: \"kubernetes.io/projected/5f5a4382-7710-48d0-a476-8464c98612dc-kube-api-access-x28cx\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kjz6s\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.267054 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kjz6s\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.368529 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kjz6s\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.368596 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kjz6s\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.368686 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x28cx\" (UniqueName: \"kubernetes.io/projected/5f5a4382-7710-48d0-a476-8464c98612dc-kube-api-access-x28cx\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kjz6s\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.374832 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kjz6s\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.376112 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kjz6s\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.390053 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x28cx\" (UniqueName: \"kubernetes.io/projected/5f5a4382-7710-48d0-a476-8464c98612dc-kube-api-access-x28cx\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-kjz6s\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:23 crc kubenswrapper[5025]: I1004 11:03:23.484983 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:03:24 crc kubenswrapper[5025]: I1004 11:03:24.019327 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s"] Oct 04 11:03:24 crc kubenswrapper[5025]: I1004 11:03:24.090352 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" event={"ID":"5f5a4382-7710-48d0-a476-8464c98612dc","Type":"ContainerStarted","Data":"248f3e22f4faa2fa2338cf108fdb209c275a431ffd6ac72fb7c3cf4404a4b947"} Oct 04 11:03:25 crc kubenswrapper[5025]: I1004 11:03:25.048598 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-9lvlk"] Oct 04 11:03:25 crc kubenswrapper[5025]: I1004 11:03:25.056948 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-bgvx5"] Oct 04 11:03:25 crc kubenswrapper[5025]: I1004 11:03:25.064532 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-9lvlk"] Oct 04 11:03:25 crc kubenswrapper[5025]: I1004 11:03:25.071302 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-bgvx5"] Oct 04 11:03:25 crc kubenswrapper[5025]: I1004 11:03:25.077811 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-t6kjx"] Oct 04 11:03:25 crc kubenswrapper[5025]: I1004 11:03:25.084448 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-t6kjx"] Oct 04 11:03:25 crc kubenswrapper[5025]: I1004 11:03:25.103919 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" event={"ID":"5f5a4382-7710-48d0-a476-8464c98612dc","Type":"ContainerStarted","Data":"fcad22bef8c91a00dbd2a512aeeef22553ee1ba95b045d330e0d1fb07275c225"} Oct 04 11:03:25 crc kubenswrapper[5025]: I1004 11:03:25.133344 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" podStartSLOduration=1.6202472970000001 podStartE2EDuration="2.133317355s" podCreationTimestamp="2025-10-04 11:03:23 +0000 UTC" firstStartedPulling="2025-10-04 11:03:24.029278606 +0000 UTC m=+1732.454245496" lastFinishedPulling="2025-10-04 11:03:24.542348674 +0000 UTC m=+1732.967315554" observedRunningTime="2025-10-04 11:03:25.123137678 +0000 UTC m=+1733.548104558" watchObservedRunningTime="2025-10-04 11:03:25.133317355 +0000 UTC m=+1733.558284275" Oct 04 11:03:26 crc kubenswrapper[5025]: I1004 11:03:26.425200 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64a21479-f44b-49c1-bc9e-fafd77a4cf64" path="/var/lib/kubelet/pods/64a21479-f44b-49c1-bc9e-fafd77a4cf64/volumes" Oct 04 11:03:26 crc kubenswrapper[5025]: I1004 11:03:26.426444 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80e8d612-99b0-4396-a8dd-53b8de8956ba" path="/var/lib/kubelet/pods/80e8d612-99b0-4396-a8dd-53b8de8956ba/volumes" Oct 04 11:03:26 crc kubenswrapper[5025]: I1004 11:03:26.427586 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebdb7464-8cf2-439a-8789-e8b7617517fc" path="/var/lib/kubelet/pods/ebdb7464-8cf2-439a-8789-e8b7617517fc/volumes" Oct 04 11:03:35 crc kubenswrapper[5025]: I1004 11:03:35.413135 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:03:35 crc kubenswrapper[5025]: E1004 11:03:35.413820 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:03:36 crc kubenswrapper[5025]: I1004 11:03:36.037821 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-4f45-account-create-xmrr8"] Oct 04 11:03:36 crc kubenswrapper[5025]: I1004 11:03:36.049924 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-55f2-account-create-n2tgb"] Oct 04 11:03:36 crc kubenswrapper[5025]: I1004 11:03:36.059229 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-4f45-account-create-xmrr8"] Oct 04 11:03:36 crc kubenswrapper[5025]: I1004 11:03:36.066936 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-6551-account-create-bwsv6"] Oct 04 11:03:36 crc kubenswrapper[5025]: I1004 11:03:36.073789 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-55f2-account-create-n2tgb"] Oct 04 11:03:36 crc kubenswrapper[5025]: I1004 11:03:36.079943 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-6551-account-create-bwsv6"] Oct 04 11:03:36 crc kubenswrapper[5025]: I1004 11:03:36.434279 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24" path="/var/lib/kubelet/pods/0bac31dc-c26e-4e8b-bd2d-19e4b4cb9f24/volumes" Oct 04 11:03:36 crc kubenswrapper[5025]: I1004 11:03:36.434998 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="603ede55-aaba-46e1-80fc-ba3ee57efa10" path="/var/lib/kubelet/pods/603ede55-aaba-46e1-80fc-ba3ee57efa10/volumes" Oct 04 11:03:36 crc kubenswrapper[5025]: I1004 11:03:36.435724 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7322f1a7-deea-4d50-a9c9-5cff160a4cbb" path="/var/lib/kubelet/pods/7322f1a7-deea-4d50-a9c9-5cff160a4cbb/volumes" Oct 04 11:03:47 crc kubenswrapper[5025]: I1004 11:03:47.411248 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:03:47 crc kubenswrapper[5025]: E1004 11:03:47.412109 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:03:48 crc kubenswrapper[5025]: I1004 11:03:48.696688 5025 scope.go:117] "RemoveContainer" containerID="015d8ba1eef6ebe0ebc1c43f6e6e47bf6cba242f756777f201743f97872fc491" Oct 04 11:03:48 crc kubenswrapper[5025]: I1004 11:03:48.726638 5025 scope.go:117] "RemoveContainer" containerID="0a25d1cb9587b5420387869af2bc1fa857e9c0c55125ac33f569b2ce792ced10" Oct 04 11:03:48 crc kubenswrapper[5025]: I1004 11:03:48.804137 5025 scope.go:117] "RemoveContainer" containerID="32fefd7216c7fdfb5e4b657a7a7bd7a88d0a166619de4323b1694234028874cb" Oct 04 11:03:48 crc kubenswrapper[5025]: I1004 11:03:48.847580 5025 scope.go:117] "RemoveContainer" containerID="119283c99064e440b346a12cc67ff3aa47f7ec86d892751d424358afe9eeaf5c" Oct 04 11:03:48 crc kubenswrapper[5025]: I1004 11:03:48.883159 5025 scope.go:117] "RemoveContainer" containerID="bc5c258a101b58a8e965b16b9856cad0d3faa7e793363769df66f4228fa6f4ea" Oct 04 11:03:48 crc kubenswrapper[5025]: I1004 11:03:48.924214 5025 scope.go:117] "RemoveContainer" containerID="599e47f8c41209262f3f01ba560a31130c548fa284490bdd451e9a6e8cfd336c" Oct 04 11:03:59 crc kubenswrapper[5025]: I1004 11:03:59.046352 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fdcd7"] Oct 04 11:03:59 crc kubenswrapper[5025]: I1004 11:03:59.057376 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fdcd7"] Oct 04 11:04:00 crc kubenswrapper[5025]: I1004 11:04:00.411125 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:04:00 crc kubenswrapper[5025]: E1004 11:04:00.411659 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:04:00 crc kubenswrapper[5025]: I1004 11:04:00.424112 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70c3c42b-9625-43a5-9d21-a145d8380fc0" path="/var/lib/kubelet/pods/70c3c42b-9625-43a5-9d21-a145d8380fc0/volumes" Oct 04 11:04:03 crc kubenswrapper[5025]: I1004 11:04:03.482794 5025 generic.go:334] "Generic (PLEG): container finished" podID="5f5a4382-7710-48d0-a476-8464c98612dc" containerID="fcad22bef8c91a00dbd2a512aeeef22553ee1ba95b045d330e0d1fb07275c225" exitCode=0 Oct 04 11:04:03 crc kubenswrapper[5025]: I1004 11:04:03.482907 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" event={"ID":"5f5a4382-7710-48d0-a476-8464c98612dc","Type":"ContainerDied","Data":"fcad22bef8c91a00dbd2a512aeeef22553ee1ba95b045d330e0d1fb07275c225"} Oct 04 11:04:04 crc kubenswrapper[5025]: I1004 11:04:04.908383 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:04:04 crc kubenswrapper[5025]: I1004 11:04:04.980302 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-inventory\") pod \"5f5a4382-7710-48d0-a476-8464c98612dc\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " Oct 04 11:04:04 crc kubenswrapper[5025]: I1004 11:04:04.980522 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-ssh-key\") pod \"5f5a4382-7710-48d0-a476-8464c98612dc\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " Oct 04 11:04:04 crc kubenswrapper[5025]: I1004 11:04:04.980609 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x28cx\" (UniqueName: \"kubernetes.io/projected/5f5a4382-7710-48d0-a476-8464c98612dc-kube-api-access-x28cx\") pod \"5f5a4382-7710-48d0-a476-8464c98612dc\" (UID: \"5f5a4382-7710-48d0-a476-8464c98612dc\") " Oct 04 11:04:04 crc kubenswrapper[5025]: I1004 11:04:04.985469 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f5a4382-7710-48d0-a476-8464c98612dc-kube-api-access-x28cx" (OuterVolumeSpecName: "kube-api-access-x28cx") pod "5f5a4382-7710-48d0-a476-8464c98612dc" (UID: "5f5a4382-7710-48d0-a476-8464c98612dc"). InnerVolumeSpecName "kube-api-access-x28cx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.005651 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5f5a4382-7710-48d0-a476-8464c98612dc" (UID: "5f5a4382-7710-48d0-a476-8464c98612dc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.009671 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-inventory" (OuterVolumeSpecName: "inventory") pod "5f5a4382-7710-48d0-a476-8464c98612dc" (UID: "5f5a4382-7710-48d0-a476-8464c98612dc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.083753 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.083802 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x28cx\" (UniqueName: \"kubernetes.io/projected/5f5a4382-7710-48d0-a476-8464c98612dc-kube-api-access-x28cx\") on node \"crc\" DevicePath \"\"" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.083822 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f5a4382-7710-48d0-a476-8464c98612dc-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.502244 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" event={"ID":"5f5a4382-7710-48d0-a476-8464c98612dc","Type":"ContainerDied","Data":"248f3e22f4faa2fa2338cf108fdb209c275a431ffd6ac72fb7c3cf4404a4b947"} Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.502289 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="248f3e22f4faa2fa2338cf108fdb209c275a431ffd6ac72fb7c3cf4404a4b947" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.502304 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-kjz6s" Oct 04 11:04:05 crc kubenswrapper[5025]: E1004 11:04:05.583576 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f5a4382_7710_48d0_a476_8464c98612dc.slice\": RecentStats: unable to find data in memory cache]" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.600048 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl"] Oct 04 11:04:05 crc kubenswrapper[5025]: E1004 11:04:05.600541 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f5a4382-7710-48d0-a476-8464c98612dc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.600569 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f5a4382-7710-48d0-a476-8464c98612dc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.600808 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f5a4382-7710-48d0-a476-8464c98612dc" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.601623 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.605511 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.605716 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.606032 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.606682 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.611251 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl"] Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.692971 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkp2v\" (UniqueName: \"kubernetes.io/projected/015bb9a3-af1a-4721-930b-df387aefb2b8-kube-api-access-lkp2v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.693140 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.693190 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.794799 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkp2v\" (UniqueName: \"kubernetes.io/projected/015bb9a3-af1a-4721-930b-df387aefb2b8-kube-api-access-lkp2v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.794889 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.794922 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.814801 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.817560 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.819831 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkp2v\" (UniqueName: \"kubernetes.io/projected/015bb9a3-af1a-4721-930b-df387aefb2b8-kube-api-access-lkp2v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:05 crc kubenswrapper[5025]: I1004 11:04:05.922402 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:04:06 crc kubenswrapper[5025]: I1004 11:04:06.472631 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl"] Oct 04 11:04:06 crc kubenswrapper[5025]: I1004 11:04:06.513756 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" event={"ID":"015bb9a3-af1a-4721-930b-df387aefb2b8","Type":"ContainerStarted","Data":"fdc91c98d608bd51041bd8450748e84ecd9cfea62a1c8ac6e11336ccd281db73"} Oct 04 11:04:07 crc kubenswrapper[5025]: I1004 11:04:07.522864 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" event={"ID":"015bb9a3-af1a-4721-930b-df387aefb2b8","Type":"ContainerStarted","Data":"d3f56626b171acb7ce16e34027eb5184357f7d60c80e5a7f972e6cef2fd252a7"} Oct 04 11:04:07 crc kubenswrapper[5025]: I1004 11:04:07.536062 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" podStartSLOduration=2.031108357 podStartE2EDuration="2.536041297s" podCreationTimestamp="2025-10-04 11:04:05 +0000 UTC" firstStartedPulling="2025-10-04 11:04:06.480839634 +0000 UTC m=+1774.905806534" lastFinishedPulling="2025-10-04 11:04:06.985772574 +0000 UTC m=+1775.410739474" observedRunningTime="2025-10-04 11:04:07.534912554 +0000 UTC m=+1775.959879434" watchObservedRunningTime="2025-10-04 11:04:07.536041297 +0000 UTC m=+1775.961008177" Oct 04 11:04:13 crc kubenswrapper[5025]: I1004 11:04:13.411177 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:04:13 crc kubenswrapper[5025]: E1004 11:04:13.412229 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:04:22 crc kubenswrapper[5025]: I1004 11:04:22.045125 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-tpdlf"] Oct 04 11:04:22 crc kubenswrapper[5025]: I1004 11:04:22.052357 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wdjvg"] Oct 04 11:04:22 crc kubenswrapper[5025]: I1004 11:04:22.060130 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-tpdlf"] Oct 04 11:04:22 crc kubenswrapper[5025]: I1004 11:04:22.068149 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-wdjvg"] Oct 04 11:04:22 crc kubenswrapper[5025]: I1004 11:04:22.423500 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03d77cd8-31b2-4500-92c3-55ca3c0d09de" path="/var/lib/kubelet/pods/03d77cd8-31b2-4500-92c3-55ca3c0d09de/volumes" Oct 04 11:04:22 crc kubenswrapper[5025]: I1004 11:04:22.424298 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57e9191a-aa60-498f-b621-2b83a7b795ed" path="/var/lib/kubelet/pods/57e9191a-aa60-498f-b621-2b83a7b795ed/volumes" Oct 04 11:04:28 crc kubenswrapper[5025]: I1004 11:04:28.411392 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:04:28 crc kubenswrapper[5025]: E1004 11:04:28.412547 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:04:41 crc kubenswrapper[5025]: I1004 11:04:41.411644 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:04:41 crc kubenswrapper[5025]: E1004 11:04:41.412340 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:04:49 crc kubenswrapper[5025]: I1004 11:04:49.073241 5025 scope.go:117] "RemoveContainer" containerID="cc4e93f993fe836c8759233ae9ab35fee56e949b8b2916fa754a0f4196d7780d" Oct 04 11:04:49 crc kubenswrapper[5025]: I1004 11:04:49.130907 5025 scope.go:117] "RemoveContainer" containerID="10e5e89d8d185b91a6ff58dbe1360f1343045faa872a02142d18c5b1cb774ab6" Oct 04 11:04:49 crc kubenswrapper[5025]: I1004 11:04:49.188917 5025 scope.go:117] "RemoveContainer" containerID="ce3d995cbaac27290050e3da1c38213b610291d8960d46702cf5decf6093bd1b" Oct 04 11:04:55 crc kubenswrapper[5025]: I1004 11:04:55.411688 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:04:55 crc kubenswrapper[5025]: E1004 11:04:55.412489 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:05:04 crc kubenswrapper[5025]: I1004 11:05:04.085443 5025 generic.go:334] "Generic (PLEG): container finished" podID="015bb9a3-af1a-4721-930b-df387aefb2b8" containerID="d3f56626b171acb7ce16e34027eb5184357f7d60c80e5a7f972e6cef2fd252a7" exitCode=2 Oct 04 11:05:04 crc kubenswrapper[5025]: I1004 11:05:04.085579 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" event={"ID":"015bb9a3-af1a-4721-930b-df387aefb2b8","Type":"ContainerDied","Data":"d3f56626b171acb7ce16e34027eb5184357f7d60c80e5a7f972e6cef2fd252a7"} Oct 04 11:05:05 crc kubenswrapper[5025]: I1004 11:05:05.489261 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:05:05 crc kubenswrapper[5025]: I1004 11:05:05.576465 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkp2v\" (UniqueName: \"kubernetes.io/projected/015bb9a3-af1a-4721-930b-df387aefb2b8-kube-api-access-lkp2v\") pod \"015bb9a3-af1a-4721-930b-df387aefb2b8\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " Oct 04 11:05:05 crc kubenswrapper[5025]: I1004 11:05:05.576579 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-ssh-key\") pod \"015bb9a3-af1a-4721-930b-df387aefb2b8\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " Oct 04 11:05:05 crc kubenswrapper[5025]: I1004 11:05:05.576854 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-inventory\") pod \"015bb9a3-af1a-4721-930b-df387aefb2b8\" (UID: \"015bb9a3-af1a-4721-930b-df387aefb2b8\") " Oct 04 11:05:05 crc kubenswrapper[5025]: I1004 11:05:05.584124 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/015bb9a3-af1a-4721-930b-df387aefb2b8-kube-api-access-lkp2v" (OuterVolumeSpecName: "kube-api-access-lkp2v") pod "015bb9a3-af1a-4721-930b-df387aefb2b8" (UID: "015bb9a3-af1a-4721-930b-df387aefb2b8"). InnerVolumeSpecName "kube-api-access-lkp2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:05:05 crc kubenswrapper[5025]: I1004 11:05:05.612621 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-inventory" (OuterVolumeSpecName: "inventory") pod "015bb9a3-af1a-4721-930b-df387aefb2b8" (UID: "015bb9a3-af1a-4721-930b-df387aefb2b8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:05:05 crc kubenswrapper[5025]: I1004 11:05:05.615067 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "015bb9a3-af1a-4721-930b-df387aefb2b8" (UID: "015bb9a3-af1a-4721-930b-df387aefb2b8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:05:05 crc kubenswrapper[5025]: I1004 11:05:05.679603 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:05 crc kubenswrapper[5025]: I1004 11:05:05.679654 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkp2v\" (UniqueName: \"kubernetes.io/projected/015bb9a3-af1a-4721-930b-df387aefb2b8-kube-api-access-lkp2v\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:05 crc kubenswrapper[5025]: I1004 11:05:05.679678 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/015bb9a3-af1a-4721-930b-df387aefb2b8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:05:06 crc kubenswrapper[5025]: I1004 11:05:06.106882 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" event={"ID":"015bb9a3-af1a-4721-930b-df387aefb2b8","Type":"ContainerDied","Data":"fdc91c98d608bd51041bd8450748e84ecd9cfea62a1c8ac6e11336ccd281db73"} Oct 04 11:05:06 crc kubenswrapper[5025]: I1004 11:05:06.106929 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdc91c98d608bd51041bd8450748e84ecd9cfea62a1c8ac6e11336ccd281db73" Oct 04 11:05:06 crc kubenswrapper[5025]: I1004 11:05:06.106933 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl" Oct 04 11:05:08 crc kubenswrapper[5025]: I1004 11:05:08.038471 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-bs82b"] Oct 04 11:05:08 crc kubenswrapper[5025]: I1004 11:05:08.044830 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-bs82b"] Oct 04 11:05:08 crc kubenswrapper[5025]: I1004 11:05:08.412051 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:05:08 crc kubenswrapper[5025]: E1004 11:05:08.412352 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:05:08 crc kubenswrapper[5025]: I1004 11:05:08.424091 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65baca16-d04e-44ea-8b1b-0f573bbd0d21" path="/var/lib/kubelet/pods/65baca16-d04e-44ea-8b1b-0f573bbd0d21/volumes" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.037831 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5"] Oct 04 11:05:13 crc kubenswrapper[5025]: E1004 11:05:13.038998 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="015bb9a3-af1a-4721-930b-df387aefb2b8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.039045 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="015bb9a3-af1a-4721-930b-df387aefb2b8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.039433 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="015bb9a3-af1a-4721-930b-df387aefb2b8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.040355 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.047751 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.048153 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.048187 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.048149 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5"] Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.048718 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.108454 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwllx\" (UniqueName: \"kubernetes.io/projected/7f1f31a9-c086-48bb-86e3-6db0b697de2e-kube-api-access-bwllx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.108813 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.108872 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.210562 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwllx\" (UniqueName: \"kubernetes.io/projected/7f1f31a9-c086-48bb-86e3-6db0b697de2e-kube-api-access-bwllx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.210666 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.210724 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.217045 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.217200 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.229637 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwllx\" (UniqueName: \"kubernetes.io/projected/7f1f31a9-c086-48bb-86e3-6db0b697de2e-kube-api-access-bwllx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.369648 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:05:13 crc kubenswrapper[5025]: I1004 11:05:13.723818 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5"] Oct 04 11:05:14 crc kubenswrapper[5025]: I1004 11:05:14.170995 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" event={"ID":"7f1f31a9-c086-48bb-86e3-6db0b697de2e","Type":"ContainerStarted","Data":"eeedec27fc6dc3fc3a18029606e3101224b54f8442734086823179a53d4031e3"} Oct 04 11:05:15 crc kubenswrapper[5025]: I1004 11:05:15.180224 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" event={"ID":"7f1f31a9-c086-48bb-86e3-6db0b697de2e","Type":"ContainerStarted","Data":"73ead93dcbb502aa462d17c6f6cdf4be4b8e0fdef90397ddfcca29b1045091b8"} Oct 04 11:05:15 crc kubenswrapper[5025]: I1004 11:05:15.206188 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" podStartSLOduration=1.618174669 podStartE2EDuration="2.206162313s" podCreationTimestamp="2025-10-04 11:05:13 +0000 UTC" firstStartedPulling="2025-10-04 11:05:13.727295472 +0000 UTC m=+1842.152262352" lastFinishedPulling="2025-10-04 11:05:14.315283106 +0000 UTC m=+1842.740249996" observedRunningTime="2025-10-04 11:05:15.198678644 +0000 UTC m=+1843.623645564" watchObservedRunningTime="2025-10-04 11:05:15.206162313 +0000 UTC m=+1843.631129203" Oct 04 11:05:20 crc kubenswrapper[5025]: I1004 11:05:20.414088 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:05:20 crc kubenswrapper[5025]: E1004 11:05:20.415241 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:05:31 crc kubenswrapper[5025]: I1004 11:05:31.411700 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:05:31 crc kubenswrapper[5025]: E1004 11:05:31.412461 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:05:44 crc kubenswrapper[5025]: I1004 11:05:44.410980 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:05:44 crc kubenswrapper[5025]: E1004 11:05:44.411910 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:05:49 crc kubenswrapper[5025]: I1004 11:05:49.312673 5025 scope.go:117] "RemoveContainer" containerID="b9838b5f5a4b67498f3df08464bb9d02bc5ce4e50e8ba8b921cf1702af0a1a3d" Oct 04 11:05:55 crc kubenswrapper[5025]: I1004 11:05:55.412515 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:05:55 crc kubenswrapper[5025]: E1004 11:05:55.413467 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:05:59 crc kubenswrapper[5025]: I1004 11:05:59.666542 5025 generic.go:334] "Generic (PLEG): container finished" podID="7f1f31a9-c086-48bb-86e3-6db0b697de2e" containerID="73ead93dcbb502aa462d17c6f6cdf4be4b8e0fdef90397ddfcca29b1045091b8" exitCode=0 Oct 04 11:05:59 crc kubenswrapper[5025]: I1004 11:05:59.666648 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" event={"ID":"7f1f31a9-c086-48bb-86e3-6db0b697de2e","Type":"ContainerDied","Data":"73ead93dcbb502aa462d17c6f6cdf4be4b8e0fdef90397ddfcca29b1045091b8"} Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.167707 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.315198 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwllx\" (UniqueName: \"kubernetes.io/projected/7f1f31a9-c086-48bb-86e3-6db0b697de2e-kube-api-access-bwllx\") pod \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.315389 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-ssh-key\") pod \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.315489 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-inventory\") pod \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\" (UID: \"7f1f31a9-c086-48bb-86e3-6db0b697de2e\") " Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.321945 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f1f31a9-c086-48bb-86e3-6db0b697de2e-kube-api-access-bwllx" (OuterVolumeSpecName: "kube-api-access-bwllx") pod "7f1f31a9-c086-48bb-86e3-6db0b697de2e" (UID: "7f1f31a9-c086-48bb-86e3-6db0b697de2e"). InnerVolumeSpecName "kube-api-access-bwllx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.346118 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-inventory" (OuterVolumeSpecName: "inventory") pod "7f1f31a9-c086-48bb-86e3-6db0b697de2e" (UID: "7f1f31a9-c086-48bb-86e3-6db0b697de2e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.350143 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7f1f31a9-c086-48bb-86e3-6db0b697de2e" (UID: "7f1f31a9-c086-48bb-86e3-6db0b697de2e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.417762 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.417797 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7f1f31a9-c086-48bb-86e3-6db0b697de2e-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.417807 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwllx\" (UniqueName: \"kubernetes.io/projected/7f1f31a9-c086-48bb-86e3-6db0b697de2e-kube-api-access-bwllx\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.689808 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" event={"ID":"7f1f31a9-c086-48bb-86e3-6db0b697de2e","Type":"ContainerDied","Data":"eeedec27fc6dc3fc3a18029606e3101224b54f8442734086823179a53d4031e3"} Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.690275 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eeedec27fc6dc3fc3a18029606e3101224b54f8442734086823179a53d4031e3" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.689900 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.829170 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-jv7vr"] Oct 04 11:06:01 crc kubenswrapper[5025]: E1004 11:06:01.830007 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f1f31a9-c086-48bb-86e3-6db0b697de2e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.830088 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f1f31a9-c086-48bb-86e3-6db0b697de2e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.830676 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f1f31a9-c086-48bb-86e3-6db0b697de2e" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.832359 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.838258 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-jv7vr"] Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.839283 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.839662 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.840405 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.840776 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.927637 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-jv7vr\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.928149 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-jv7vr\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:01 crc kubenswrapper[5025]: I1004 11:06:01.928489 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clwhm\" (UniqueName: \"kubernetes.io/projected/e4ba226b-2376-4bd2-823c-c8fc00c46369-kube-api-access-clwhm\") pod \"ssh-known-hosts-edpm-deployment-jv7vr\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:02 crc kubenswrapper[5025]: I1004 11:06:02.030047 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-jv7vr\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:02 crc kubenswrapper[5025]: I1004 11:06:02.030161 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clwhm\" (UniqueName: \"kubernetes.io/projected/e4ba226b-2376-4bd2-823c-c8fc00c46369-kube-api-access-clwhm\") pod \"ssh-known-hosts-edpm-deployment-jv7vr\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:02 crc kubenswrapper[5025]: I1004 11:06:02.030231 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-jv7vr\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:02 crc kubenswrapper[5025]: I1004 11:06:02.035794 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-jv7vr\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:02 crc kubenswrapper[5025]: I1004 11:06:02.042775 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-jv7vr\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:02 crc kubenswrapper[5025]: I1004 11:06:02.046484 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clwhm\" (UniqueName: \"kubernetes.io/projected/e4ba226b-2376-4bd2-823c-c8fc00c46369-kube-api-access-clwhm\") pod \"ssh-known-hosts-edpm-deployment-jv7vr\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:02 crc kubenswrapper[5025]: I1004 11:06:02.155686 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:02 crc kubenswrapper[5025]: I1004 11:06:02.700242 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-jv7vr"] Oct 04 11:06:02 crc kubenswrapper[5025]: W1004 11:06:02.712497 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4ba226b_2376_4bd2_823c_c8fc00c46369.slice/crio-0093d1832f79906a5178ca6275a32d5b4a42f7036b50a4600d0ffdb169906eaa WatchSource:0}: Error finding container 0093d1832f79906a5178ca6275a32d5b4a42f7036b50a4600d0ffdb169906eaa: Status 404 returned error can't find the container with id 0093d1832f79906a5178ca6275a32d5b4a42f7036b50a4600d0ffdb169906eaa Oct 04 11:06:03 crc kubenswrapper[5025]: I1004 11:06:03.711917 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" event={"ID":"e4ba226b-2376-4bd2-823c-c8fc00c46369","Type":"ContainerStarted","Data":"c773576784d18ff86356db48a0fea573d95566a6e526c188dd2685f7ffbcaaca"} Oct 04 11:06:03 crc kubenswrapper[5025]: I1004 11:06:03.712491 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" event={"ID":"e4ba226b-2376-4bd2-823c-c8fc00c46369","Type":"ContainerStarted","Data":"0093d1832f79906a5178ca6275a32d5b4a42f7036b50a4600d0ffdb169906eaa"} Oct 04 11:06:03 crc kubenswrapper[5025]: I1004 11:06:03.743889 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" podStartSLOduration=2.243956146 podStartE2EDuration="2.74386622s" podCreationTimestamp="2025-10-04 11:06:01 +0000 UTC" firstStartedPulling="2025-10-04 11:06:02.715216991 +0000 UTC m=+1891.140183871" lastFinishedPulling="2025-10-04 11:06:03.215127075 +0000 UTC m=+1891.640093945" observedRunningTime="2025-10-04 11:06:03.730295184 +0000 UTC m=+1892.155262144" watchObservedRunningTime="2025-10-04 11:06:03.74386622 +0000 UTC m=+1892.168833110" Oct 04 11:06:09 crc kubenswrapper[5025]: I1004 11:06:09.411700 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:06:09 crc kubenswrapper[5025]: E1004 11:06:09.412586 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:06:10 crc kubenswrapper[5025]: I1004 11:06:10.781507 5025 generic.go:334] "Generic (PLEG): container finished" podID="e4ba226b-2376-4bd2-823c-c8fc00c46369" containerID="c773576784d18ff86356db48a0fea573d95566a6e526c188dd2685f7ffbcaaca" exitCode=0 Oct 04 11:06:10 crc kubenswrapper[5025]: I1004 11:06:10.781557 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" event={"ID":"e4ba226b-2376-4bd2-823c-c8fc00c46369","Type":"ContainerDied","Data":"c773576784d18ff86356db48a0fea573d95566a6e526c188dd2685f7ffbcaaca"} Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.224814 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.363226 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clwhm\" (UniqueName: \"kubernetes.io/projected/e4ba226b-2376-4bd2-823c-c8fc00c46369-kube-api-access-clwhm\") pod \"e4ba226b-2376-4bd2-823c-c8fc00c46369\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.363503 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-inventory-0\") pod \"e4ba226b-2376-4bd2-823c-c8fc00c46369\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.363562 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-ssh-key-openstack-edpm-ipam\") pod \"e4ba226b-2376-4bd2-823c-c8fc00c46369\" (UID: \"e4ba226b-2376-4bd2-823c-c8fc00c46369\") " Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.369604 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4ba226b-2376-4bd2-823c-c8fc00c46369-kube-api-access-clwhm" (OuterVolumeSpecName: "kube-api-access-clwhm") pod "e4ba226b-2376-4bd2-823c-c8fc00c46369" (UID: "e4ba226b-2376-4bd2-823c-c8fc00c46369"). InnerVolumeSpecName "kube-api-access-clwhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.390734 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e4ba226b-2376-4bd2-823c-c8fc00c46369" (UID: "e4ba226b-2376-4bd2-823c-c8fc00c46369"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.391543 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e4ba226b-2376-4bd2-823c-c8fc00c46369" (UID: "e4ba226b-2376-4bd2-823c-c8fc00c46369"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.466063 5025 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.466092 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e4ba226b-2376-4bd2-823c-c8fc00c46369-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.466104 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clwhm\" (UniqueName: \"kubernetes.io/projected/e4ba226b-2376-4bd2-823c-c8fc00c46369-kube-api-access-clwhm\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.810364 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" event={"ID":"e4ba226b-2376-4bd2-823c-c8fc00c46369","Type":"ContainerDied","Data":"0093d1832f79906a5178ca6275a32d5b4a42f7036b50a4600d0ffdb169906eaa"} Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.810850 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0093d1832f79906a5178ca6275a32d5b4a42f7036b50a4600d0ffdb169906eaa" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.810433 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-jv7vr" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.880570 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb"] Oct 04 11:06:12 crc kubenswrapper[5025]: E1004 11:06:12.880943 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4ba226b-2376-4bd2-823c-c8fc00c46369" containerName="ssh-known-hosts-edpm-deployment" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.880963 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4ba226b-2376-4bd2-823c-c8fc00c46369" containerName="ssh-known-hosts-edpm-deployment" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.881291 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4ba226b-2376-4bd2-823c-c8fc00c46369" containerName="ssh-known-hosts-edpm-deployment" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.881869 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.885552 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.885934 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.886398 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.886784 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.902686 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb"] Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.976644 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqvtj\" (UniqueName: \"kubernetes.io/projected/f677e673-14c1-4f54-b7db-15b99f3922d2-kube-api-access-zqvtj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g8clb\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.976920 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g8clb\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:12 crc kubenswrapper[5025]: I1004 11:06:12.977177 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g8clb\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:13 crc kubenswrapper[5025]: I1004 11:06:13.079303 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqvtj\" (UniqueName: \"kubernetes.io/projected/f677e673-14c1-4f54-b7db-15b99f3922d2-kube-api-access-zqvtj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g8clb\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:13 crc kubenswrapper[5025]: I1004 11:06:13.079457 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g8clb\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:13 crc kubenswrapper[5025]: I1004 11:06:13.079504 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g8clb\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:13 crc kubenswrapper[5025]: I1004 11:06:13.084738 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g8clb\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:13 crc kubenswrapper[5025]: I1004 11:06:13.085571 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g8clb\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:13 crc kubenswrapper[5025]: I1004 11:06:13.106787 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqvtj\" (UniqueName: \"kubernetes.io/projected/f677e673-14c1-4f54-b7db-15b99f3922d2-kube-api-access-zqvtj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-g8clb\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:13 crc kubenswrapper[5025]: I1004 11:06:13.217563 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:13 crc kubenswrapper[5025]: I1004 11:06:13.573980 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb"] Oct 04 11:06:13 crc kubenswrapper[5025]: I1004 11:06:13.828864 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" event={"ID":"f677e673-14c1-4f54-b7db-15b99f3922d2","Type":"ContainerStarted","Data":"be93a955ce43c4b8968540f241b35103e1e0d2ab8ab9f02fb07093789e4366f0"} Oct 04 11:06:14 crc kubenswrapper[5025]: I1004 11:06:14.850893 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" event={"ID":"f677e673-14c1-4f54-b7db-15b99f3922d2","Type":"ContainerStarted","Data":"9ca9fb7025d019c2ebcfa10be0504bbcf18b68f84bd2847f52491eae8a6f3155"} Oct 04 11:06:14 crc kubenswrapper[5025]: I1004 11:06:14.883196 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" podStartSLOduration=2.43282005 podStartE2EDuration="2.883177537s" podCreationTimestamp="2025-10-04 11:06:12 +0000 UTC" firstStartedPulling="2025-10-04 11:06:13.579977534 +0000 UTC m=+1902.004944414" lastFinishedPulling="2025-10-04 11:06:14.030335011 +0000 UTC m=+1902.455301901" observedRunningTime="2025-10-04 11:06:14.875640217 +0000 UTC m=+1903.300607137" watchObservedRunningTime="2025-10-04 11:06:14.883177537 +0000 UTC m=+1903.308144427" Oct 04 11:06:22 crc kubenswrapper[5025]: I1004 11:06:22.418552 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:06:22 crc kubenswrapper[5025]: E1004 11:06:22.419478 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:06:22 crc kubenswrapper[5025]: I1004 11:06:22.940653 5025 generic.go:334] "Generic (PLEG): container finished" podID="f677e673-14c1-4f54-b7db-15b99f3922d2" containerID="9ca9fb7025d019c2ebcfa10be0504bbcf18b68f84bd2847f52491eae8a6f3155" exitCode=0 Oct 04 11:06:22 crc kubenswrapper[5025]: I1004 11:06:22.940713 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" event={"ID":"f677e673-14c1-4f54-b7db-15b99f3922d2","Type":"ContainerDied","Data":"9ca9fb7025d019c2ebcfa10be0504bbcf18b68f84bd2847f52491eae8a6f3155"} Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.483790 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.527484 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-inventory\") pod \"f677e673-14c1-4f54-b7db-15b99f3922d2\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.527617 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-ssh-key\") pod \"f677e673-14c1-4f54-b7db-15b99f3922d2\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.527699 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqvtj\" (UniqueName: \"kubernetes.io/projected/f677e673-14c1-4f54-b7db-15b99f3922d2-kube-api-access-zqvtj\") pod \"f677e673-14c1-4f54-b7db-15b99f3922d2\" (UID: \"f677e673-14c1-4f54-b7db-15b99f3922d2\") " Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.556678 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f677e673-14c1-4f54-b7db-15b99f3922d2-kube-api-access-zqvtj" (OuterVolumeSpecName: "kube-api-access-zqvtj") pod "f677e673-14c1-4f54-b7db-15b99f3922d2" (UID: "f677e673-14c1-4f54-b7db-15b99f3922d2"). InnerVolumeSpecName "kube-api-access-zqvtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.564122 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-inventory" (OuterVolumeSpecName: "inventory") pod "f677e673-14c1-4f54-b7db-15b99f3922d2" (UID: "f677e673-14c1-4f54-b7db-15b99f3922d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.569251 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f677e673-14c1-4f54-b7db-15b99f3922d2" (UID: "f677e673-14c1-4f54-b7db-15b99f3922d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.629642 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.629676 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqvtj\" (UniqueName: \"kubernetes.io/projected/f677e673-14c1-4f54-b7db-15b99f3922d2-kube-api-access-zqvtj\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.629687 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f677e673-14c1-4f54-b7db-15b99f3922d2-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.967509 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" event={"ID":"f677e673-14c1-4f54-b7db-15b99f3922d2","Type":"ContainerDied","Data":"be93a955ce43c4b8968540f241b35103e1e0d2ab8ab9f02fb07093789e4366f0"} Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.967564 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be93a955ce43c4b8968540f241b35103e1e0d2ab8ab9f02fb07093789e4366f0" Oct 04 11:06:24 crc kubenswrapper[5025]: I1004 11:06:24.967579 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-g8clb" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.062090 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5"] Oct 04 11:06:25 crc kubenswrapper[5025]: E1004 11:06:25.062564 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f677e673-14c1-4f54-b7db-15b99f3922d2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.062588 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="f677e673-14c1-4f54-b7db-15b99f3922d2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.062916 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="f677e673-14c1-4f54-b7db-15b99f3922d2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.064257 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.073188 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.073894 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5"] Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.078539 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.079663 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.107429 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.139119 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.139239 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.139315 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfdbn\" (UniqueName: \"kubernetes.io/projected/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-kube-api-access-gfdbn\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.241140 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.241242 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.241315 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfdbn\" (UniqueName: \"kubernetes.io/projected/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-kube-api-access-gfdbn\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.249786 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.249805 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.261377 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfdbn\" (UniqueName: \"kubernetes.io/projected/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-kube-api-access-gfdbn\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.389100 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:25 crc kubenswrapper[5025]: I1004 11:06:25.957566 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5"] Oct 04 11:06:26 crc kubenswrapper[5025]: I1004 11:06:26.993514 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" event={"ID":"c9ecdd38-1c7f-4739-b9e5-80ef801564d9","Type":"ContainerStarted","Data":"d3b66d2f81fa757554c86efda081e5564bdfc07df4d85abb2e397ed76d720985"} Oct 04 11:06:26 crc kubenswrapper[5025]: I1004 11:06:26.994763 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" event={"ID":"c9ecdd38-1c7f-4739-b9e5-80ef801564d9","Type":"ContainerStarted","Data":"e4386b0f346657a96b579f9a4dd4f85892034d56e6c8d4bd139969a5bf171e96"} Oct 04 11:06:27 crc kubenswrapper[5025]: I1004 11:06:27.021006 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" podStartSLOduration=1.636082126 podStartE2EDuration="2.020982722s" podCreationTimestamp="2025-10-04 11:06:25 +0000 UTC" firstStartedPulling="2025-10-04 11:06:25.976496701 +0000 UTC m=+1914.401463601" lastFinishedPulling="2025-10-04 11:06:26.361397307 +0000 UTC m=+1914.786364197" observedRunningTime="2025-10-04 11:06:27.01648063 +0000 UTC m=+1915.441447570" watchObservedRunningTime="2025-10-04 11:06:27.020982722 +0000 UTC m=+1915.445949612" Oct 04 11:06:36 crc kubenswrapper[5025]: I1004 11:06:36.094865 5025 generic.go:334] "Generic (PLEG): container finished" podID="c9ecdd38-1c7f-4739-b9e5-80ef801564d9" containerID="d3b66d2f81fa757554c86efda081e5564bdfc07df4d85abb2e397ed76d720985" exitCode=0 Oct 04 11:06:36 crc kubenswrapper[5025]: I1004 11:06:36.094933 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" event={"ID":"c9ecdd38-1c7f-4739-b9e5-80ef801564d9","Type":"ContainerDied","Data":"d3b66d2f81fa757554c86efda081e5564bdfc07df4d85abb2e397ed76d720985"} Oct 04 11:06:36 crc kubenswrapper[5025]: I1004 11:06:36.411521 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:06:36 crc kubenswrapper[5025]: E1004 11:06:36.412108 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:06:37 crc kubenswrapper[5025]: I1004 11:06:37.556049 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:37 crc kubenswrapper[5025]: I1004 11:06:37.630750 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gfdbn\" (UniqueName: \"kubernetes.io/projected/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-kube-api-access-gfdbn\") pod \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " Oct 04 11:06:37 crc kubenswrapper[5025]: I1004 11:06:37.630814 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-ssh-key\") pod \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " Oct 04 11:06:37 crc kubenswrapper[5025]: I1004 11:06:37.631053 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-inventory\") pod \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\" (UID: \"c9ecdd38-1c7f-4739-b9e5-80ef801564d9\") " Oct 04 11:06:37 crc kubenswrapper[5025]: I1004 11:06:37.636349 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-kube-api-access-gfdbn" (OuterVolumeSpecName: "kube-api-access-gfdbn") pod "c9ecdd38-1c7f-4739-b9e5-80ef801564d9" (UID: "c9ecdd38-1c7f-4739-b9e5-80ef801564d9"). InnerVolumeSpecName "kube-api-access-gfdbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:06:37 crc kubenswrapper[5025]: I1004 11:06:37.657929 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-inventory" (OuterVolumeSpecName: "inventory") pod "c9ecdd38-1c7f-4739-b9e5-80ef801564d9" (UID: "c9ecdd38-1c7f-4739-b9e5-80ef801564d9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:37 crc kubenswrapper[5025]: I1004 11:06:37.665038 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c9ecdd38-1c7f-4739-b9e5-80ef801564d9" (UID: "c9ecdd38-1c7f-4739-b9e5-80ef801564d9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:06:37 crc kubenswrapper[5025]: I1004 11:06:37.742238 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gfdbn\" (UniqueName: \"kubernetes.io/projected/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-kube-api-access-gfdbn\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:37 crc kubenswrapper[5025]: I1004 11:06:37.742544 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:37 crc kubenswrapper[5025]: I1004 11:06:37.742731 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9ecdd38-1c7f-4739-b9e5-80ef801564d9-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.118304 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" event={"ID":"c9ecdd38-1c7f-4739-b9e5-80ef801564d9","Type":"ContainerDied","Data":"e4386b0f346657a96b579f9a4dd4f85892034d56e6c8d4bd139969a5bf171e96"} Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.118352 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4386b0f346657a96b579f9a4dd4f85892034d56e6c8d4bd139969a5bf171e96" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.118424 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.206978 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8"] Oct 04 11:06:38 crc kubenswrapper[5025]: E1004 11:06:38.207421 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ecdd38-1c7f-4739-b9e5-80ef801564d9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.207442 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ecdd38-1c7f-4739-b9e5-80ef801564d9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.207635 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9ecdd38-1c7f-4739-b9e5-80ef801564d9" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.208319 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.211084 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.211397 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.211639 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.211796 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.212005 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.212213 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.212430 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.212565 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.225981 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8"] Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.354734 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvd2q\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-kube-api-access-tvd2q\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.354774 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.354798 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.354833 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.354865 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.354884 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.354945 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.354975 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.354993 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.355045 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.355080 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.355103 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.355124 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.355237 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456443 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvd2q\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-kube-api-access-tvd2q\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456508 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456579 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456613 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456643 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456662 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456687 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456706 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456722 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456737 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456764 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456787 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456806 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.456838 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.467073 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.467370 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.468481 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.469160 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.469305 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.469889 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.470198 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.470222 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.470222 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.470481 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.470633 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.472520 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.474586 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.478228 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvd2q\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-kube-api-access-tvd2q\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:38 crc kubenswrapper[5025]: I1004 11:06:38.569626 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:06:39 crc kubenswrapper[5025]: I1004 11:06:39.149053 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8"] Oct 04 11:06:40 crc kubenswrapper[5025]: I1004 11:06:40.142492 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" event={"ID":"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9","Type":"ContainerStarted","Data":"f8728f318e95e673344a17fef2fb413901629107da839032fc8641add144d54b"} Oct 04 11:06:40 crc kubenswrapper[5025]: I1004 11:06:40.142922 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" event={"ID":"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9","Type":"ContainerStarted","Data":"43d2ba72c3eb503824bb1a8740c0116306e3e0debd26d08ec68f323b5f8cadb8"} Oct 04 11:06:40 crc kubenswrapper[5025]: I1004 11:06:40.164468 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" podStartSLOduration=1.726927712 podStartE2EDuration="2.164452374s" podCreationTimestamp="2025-10-04 11:06:38 +0000 UTC" firstStartedPulling="2025-10-04 11:06:39.160422595 +0000 UTC m=+1927.585389465" lastFinishedPulling="2025-10-04 11:06:39.597947247 +0000 UTC m=+1928.022914127" observedRunningTime="2025-10-04 11:06:40.161324493 +0000 UTC m=+1928.586291373" watchObservedRunningTime="2025-10-04 11:06:40.164452374 +0000 UTC m=+1928.589419254" Oct 04 11:06:47 crc kubenswrapper[5025]: I1004 11:06:47.411752 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:06:47 crc kubenswrapper[5025]: E1004 11:06:47.412396 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:07:01 crc kubenswrapper[5025]: I1004 11:07:01.412922 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:07:01 crc kubenswrapper[5025]: E1004 11:07:01.414348 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:07:13 crc kubenswrapper[5025]: I1004 11:07:13.411564 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:07:13 crc kubenswrapper[5025]: E1004 11:07:13.412188 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:07:19 crc kubenswrapper[5025]: I1004 11:07:19.571522 5025 generic.go:334] "Generic (PLEG): container finished" podID="2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" containerID="f8728f318e95e673344a17fef2fb413901629107da839032fc8641add144d54b" exitCode=0 Oct 04 11:07:19 crc kubenswrapper[5025]: I1004 11:07:19.571622 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" event={"ID":"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9","Type":"ContainerDied","Data":"f8728f318e95e673344a17fef2fb413901629107da839032fc8641add144d54b"} Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.051326 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164347 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-inventory\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164430 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164577 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-bootstrap-combined-ca-bundle\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164640 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164674 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ssh-key\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164706 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-libvirt-combined-ca-bundle\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164740 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-ovn-default-certs-0\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164796 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-repo-setup-combined-ca-bundle\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164838 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvd2q\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-kube-api-access-tvd2q\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164883 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ovn-combined-ca-bundle\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164960 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-telemetry-combined-ca-bundle\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.164991 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-neutron-metadata-combined-ca-bundle\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.165045 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.165076 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-nova-combined-ca-bundle\") pod \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\" (UID: \"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9\") " Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.171555 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.171807 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.172681 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.173249 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.173335 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-kube-api-access-tvd2q" (OuterVolumeSpecName: "kube-api-access-tvd2q") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "kube-api-access-tvd2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.173985 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.175947 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.176037 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.176560 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.176817 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.179387 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.199670 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.211892 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-inventory" (OuterVolumeSpecName: "inventory") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.225317 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" (UID: "2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.267775 5025 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.267830 5025 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.267857 5025 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.267880 5025 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.267901 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.267918 5025 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.267938 5025 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.267956 5025 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.267974 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.267991 5025 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.268009 5025 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.268051 5025 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.268070 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvd2q\" (UniqueName: \"kubernetes.io/projected/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-kube-api-access-tvd2q\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.268088 5025 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.598300 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" event={"ID":"2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9","Type":"ContainerDied","Data":"43d2ba72c3eb503824bb1a8740c0116306e3e0debd26d08ec68f323b5f8cadb8"} Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.598357 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43d2ba72c3eb503824bb1a8740c0116306e3e0debd26d08ec68f323b5f8cadb8" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.598385 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.757044 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d"] Oct 04 11:07:21 crc kubenswrapper[5025]: E1004 11:07:21.757648 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.757680 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.758215 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.759384 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.763913 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.770225 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.779545 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.779642 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.780367 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.785728 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d"] Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.886327 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.886391 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.886436 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6rhh\" (UniqueName: \"kubernetes.io/projected/e3d2c355-fea3-4828-8291-77d0daa90ee5-kube-api-access-h6rhh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.886546 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.886603 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.989151 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.989251 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.989417 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.989454 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.989491 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6rhh\" (UniqueName: \"kubernetes.io/projected/e3d2c355-fea3-4828-8291-77d0daa90ee5-kube-api-access-h6rhh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.990556 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.994281 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:21 crc kubenswrapper[5025]: I1004 11:07:21.995406 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:22 crc kubenswrapper[5025]: I1004 11:07:22.002639 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:22 crc kubenswrapper[5025]: I1004 11:07:22.018412 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6rhh\" (UniqueName: \"kubernetes.io/projected/e3d2c355-fea3-4828-8291-77d0daa90ee5-kube-api-access-h6rhh\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-42n5d\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:22 crc kubenswrapper[5025]: I1004 11:07:22.094608 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:07:22 crc kubenswrapper[5025]: I1004 11:07:22.701139 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d"] Oct 04 11:07:22 crc kubenswrapper[5025]: I1004 11:07:22.729596 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:07:23 crc kubenswrapper[5025]: I1004 11:07:23.620853 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" event={"ID":"e3d2c355-fea3-4828-8291-77d0daa90ee5","Type":"ContainerStarted","Data":"46311c81410e7e1e5bfd5cc597ae017b4bf7f3547c229761f04e8a011fa4689b"} Oct 04 11:07:23 crc kubenswrapper[5025]: I1004 11:07:23.620907 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" event={"ID":"e3d2c355-fea3-4828-8291-77d0daa90ee5","Type":"ContainerStarted","Data":"b50bf7cd9b7e47a883d3f29d2319383764e5c840296bff6c3c66931ecead380c"} Oct 04 11:07:23 crc kubenswrapper[5025]: I1004 11:07:23.646155 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" podStartSLOduration=2.171106419 podStartE2EDuration="2.646117074s" podCreationTimestamp="2025-10-04 11:07:21 +0000 UTC" firstStartedPulling="2025-10-04 11:07:22.729325202 +0000 UTC m=+1971.154292102" lastFinishedPulling="2025-10-04 11:07:23.204335877 +0000 UTC m=+1971.629302757" observedRunningTime="2025-10-04 11:07:23.638313806 +0000 UTC m=+1972.063280706" watchObservedRunningTime="2025-10-04 11:07:23.646117074 +0000 UTC m=+1972.071083954" Oct 04 11:07:27 crc kubenswrapper[5025]: I1004 11:07:27.412152 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:07:27 crc kubenswrapper[5025]: I1004 11:07:27.663822 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"7f9c45b9be91741a2b640d8ac9c01d5e4ddf8c93a7a041740f1c11213ec28929"} Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.527327 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tlwrl"] Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.530560 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.540623 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tlwrl"] Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.671790 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nn2h\" (UniqueName: \"kubernetes.io/projected/b4f4069d-ec58-4222-b66e-8fc60955d6ba-kube-api-access-6nn2h\") pod \"redhat-marketplace-tlwrl\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.671860 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-utilities\") pod \"redhat-marketplace-tlwrl\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.671904 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-catalog-content\") pod \"redhat-marketplace-tlwrl\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.773821 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nn2h\" (UniqueName: \"kubernetes.io/projected/b4f4069d-ec58-4222-b66e-8fc60955d6ba-kube-api-access-6nn2h\") pod \"redhat-marketplace-tlwrl\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.774162 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-utilities\") pod \"redhat-marketplace-tlwrl\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.774318 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-catalog-content\") pod \"redhat-marketplace-tlwrl\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.774737 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-utilities\") pod \"redhat-marketplace-tlwrl\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.774999 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-catalog-content\") pod \"redhat-marketplace-tlwrl\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.805396 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nn2h\" (UniqueName: \"kubernetes.io/projected/b4f4069d-ec58-4222-b66e-8fc60955d6ba-kube-api-access-6nn2h\") pod \"redhat-marketplace-tlwrl\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:53 crc kubenswrapper[5025]: I1004 11:07:53.871490 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:07:54 crc kubenswrapper[5025]: I1004 11:07:54.365333 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tlwrl"] Oct 04 11:07:54 crc kubenswrapper[5025]: I1004 11:07:54.945345 5025 generic.go:334] "Generic (PLEG): container finished" podID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerID="b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425" exitCode=0 Oct 04 11:07:54 crc kubenswrapper[5025]: I1004 11:07:54.945414 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tlwrl" event={"ID":"b4f4069d-ec58-4222-b66e-8fc60955d6ba","Type":"ContainerDied","Data":"b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425"} Oct 04 11:07:54 crc kubenswrapper[5025]: I1004 11:07:54.945496 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tlwrl" event={"ID":"b4f4069d-ec58-4222-b66e-8fc60955d6ba","Type":"ContainerStarted","Data":"5318c0ddeaf6090e7d677d9b3259e8d73d718ed4767ff5b968506d33c2ac95ed"} Oct 04 11:07:55 crc kubenswrapper[5025]: I1004 11:07:55.957008 5025 generic.go:334] "Generic (PLEG): container finished" podID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerID="37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172" exitCode=0 Oct 04 11:07:55 crc kubenswrapper[5025]: I1004 11:07:55.957261 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tlwrl" event={"ID":"b4f4069d-ec58-4222-b66e-8fc60955d6ba","Type":"ContainerDied","Data":"37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172"} Oct 04 11:07:56 crc kubenswrapper[5025]: I1004 11:07:56.969182 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tlwrl" event={"ID":"b4f4069d-ec58-4222-b66e-8fc60955d6ba","Type":"ContainerStarted","Data":"1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141"} Oct 04 11:07:57 crc kubenswrapper[5025]: I1004 11:07:56.999691 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tlwrl" podStartSLOduration=2.571383054 podStartE2EDuration="3.999669057s" podCreationTimestamp="2025-10-04 11:07:53 +0000 UTC" firstStartedPulling="2025-10-04 11:07:54.950281982 +0000 UTC m=+2003.375248882" lastFinishedPulling="2025-10-04 11:07:56.378567995 +0000 UTC m=+2004.803534885" observedRunningTime="2025-10-04 11:07:56.98881356 +0000 UTC m=+2005.413780460" watchObservedRunningTime="2025-10-04 11:07:56.999669057 +0000 UTC m=+2005.424635947" Oct 04 11:08:03 crc kubenswrapper[5025]: I1004 11:08:03.872600 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:08:03 crc kubenswrapper[5025]: I1004 11:08:03.873313 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:08:03 crc kubenswrapper[5025]: I1004 11:08:03.967832 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:08:04 crc kubenswrapper[5025]: I1004 11:08:04.094626 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:08:04 crc kubenswrapper[5025]: I1004 11:08:04.206707 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tlwrl"] Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.049965 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tlwrl" podUID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerName="registry-server" containerID="cri-o://1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141" gracePeriod=2 Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.554246 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.642710 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nn2h\" (UniqueName: \"kubernetes.io/projected/b4f4069d-ec58-4222-b66e-8fc60955d6ba-kube-api-access-6nn2h\") pod \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.642854 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-catalog-content\") pod \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.642897 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-utilities\") pod \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\" (UID: \"b4f4069d-ec58-4222-b66e-8fc60955d6ba\") " Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.644039 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-utilities" (OuterVolumeSpecName: "utilities") pod "b4f4069d-ec58-4222-b66e-8fc60955d6ba" (UID: "b4f4069d-ec58-4222-b66e-8fc60955d6ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.651033 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4f4069d-ec58-4222-b66e-8fc60955d6ba-kube-api-access-6nn2h" (OuterVolumeSpecName: "kube-api-access-6nn2h") pod "b4f4069d-ec58-4222-b66e-8fc60955d6ba" (UID: "b4f4069d-ec58-4222-b66e-8fc60955d6ba"). InnerVolumeSpecName "kube-api-access-6nn2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.660589 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4f4069d-ec58-4222-b66e-8fc60955d6ba" (UID: "b4f4069d-ec58-4222-b66e-8fc60955d6ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.744944 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nn2h\" (UniqueName: \"kubernetes.io/projected/b4f4069d-ec58-4222-b66e-8fc60955d6ba-kube-api-access-6nn2h\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.744992 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:06 crc kubenswrapper[5025]: I1004 11:08:06.745005 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4f4069d-ec58-4222-b66e-8fc60955d6ba-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.065379 5025 generic.go:334] "Generic (PLEG): container finished" podID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerID="1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141" exitCode=0 Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.065474 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tlwrl" Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.065474 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tlwrl" event={"ID":"b4f4069d-ec58-4222-b66e-8fc60955d6ba","Type":"ContainerDied","Data":"1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141"} Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.066001 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tlwrl" event={"ID":"b4f4069d-ec58-4222-b66e-8fc60955d6ba","Type":"ContainerDied","Data":"5318c0ddeaf6090e7d677d9b3259e8d73d718ed4767ff5b968506d33c2ac95ed"} Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.066083 5025 scope.go:117] "RemoveContainer" containerID="1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141" Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.104201 5025 scope.go:117] "RemoveContainer" containerID="37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172" Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.138903 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tlwrl"] Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.153349 5025 scope.go:117] "RemoveContainer" containerID="b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425" Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.154602 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tlwrl"] Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.192086 5025 scope.go:117] "RemoveContainer" containerID="1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141" Oct 04 11:08:07 crc kubenswrapper[5025]: E1004 11:08:07.193191 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141\": container with ID starting with 1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141 not found: ID does not exist" containerID="1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141" Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.193497 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141"} err="failed to get container status \"1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141\": rpc error: code = NotFound desc = could not find container \"1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141\": container with ID starting with 1745571b054fbba166c71d7681597778a921395636d7d07ff2562a4282e91141 not found: ID does not exist" Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.193533 5025 scope.go:117] "RemoveContainer" containerID="37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172" Oct 04 11:08:07 crc kubenswrapper[5025]: E1004 11:08:07.194144 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172\": container with ID starting with 37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172 not found: ID does not exist" containerID="37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172" Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.194186 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172"} err="failed to get container status \"37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172\": rpc error: code = NotFound desc = could not find container \"37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172\": container with ID starting with 37d3b89ca23c8d6d18614e468a13d9bfab217b52c93ede56c8ca4cd225e14172 not found: ID does not exist" Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.194213 5025 scope.go:117] "RemoveContainer" containerID="b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425" Oct 04 11:08:07 crc kubenswrapper[5025]: E1004 11:08:07.194516 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425\": container with ID starting with b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425 not found: ID does not exist" containerID="b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425" Oct 04 11:08:07 crc kubenswrapper[5025]: I1004 11:08:07.194542 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425"} err="failed to get container status \"b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425\": rpc error: code = NotFound desc = could not find container \"b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425\": container with ID starting with b645f30f280d90f042bcc1bf8df12856f99fb8183cdca67d5b54f979e2015425 not found: ID does not exist" Oct 04 11:08:08 crc kubenswrapper[5025]: I1004 11:08:08.432709 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" path="/var/lib/kubelet/pods/b4f4069d-ec58-4222-b66e-8fc60955d6ba/volumes" Oct 04 11:08:28 crc kubenswrapper[5025]: I1004 11:08:28.298447 5025 generic.go:334] "Generic (PLEG): container finished" podID="e3d2c355-fea3-4828-8291-77d0daa90ee5" containerID="46311c81410e7e1e5bfd5cc597ae017b4bf7f3547c229761f04e8a011fa4689b" exitCode=0 Oct 04 11:08:28 crc kubenswrapper[5025]: I1004 11:08:28.298824 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" event={"ID":"e3d2c355-fea3-4828-8291-77d0daa90ee5","Type":"ContainerDied","Data":"46311c81410e7e1e5bfd5cc597ae017b4bf7f3547c229761f04e8a011fa4689b"} Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.757811 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.813059 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-inventory\") pod \"e3d2c355-fea3-4828-8291-77d0daa90ee5\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.813326 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ssh-key\") pod \"e3d2c355-fea3-4828-8291-77d0daa90ee5\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.813394 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovn-combined-ca-bundle\") pod \"e3d2c355-fea3-4828-8291-77d0daa90ee5\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.813548 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6rhh\" (UniqueName: \"kubernetes.io/projected/e3d2c355-fea3-4828-8291-77d0daa90ee5-kube-api-access-h6rhh\") pod \"e3d2c355-fea3-4828-8291-77d0daa90ee5\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.813591 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovncontroller-config-0\") pod \"e3d2c355-fea3-4828-8291-77d0daa90ee5\" (UID: \"e3d2c355-fea3-4828-8291-77d0daa90ee5\") " Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.818636 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "e3d2c355-fea3-4828-8291-77d0daa90ee5" (UID: "e3d2c355-fea3-4828-8291-77d0daa90ee5"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.821702 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3d2c355-fea3-4828-8291-77d0daa90ee5-kube-api-access-h6rhh" (OuterVolumeSpecName: "kube-api-access-h6rhh") pod "e3d2c355-fea3-4828-8291-77d0daa90ee5" (UID: "e3d2c355-fea3-4828-8291-77d0daa90ee5"). InnerVolumeSpecName "kube-api-access-h6rhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.840156 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-inventory" (OuterVolumeSpecName: "inventory") pod "e3d2c355-fea3-4828-8291-77d0daa90ee5" (UID: "e3d2c355-fea3-4828-8291-77d0daa90ee5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.844592 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "e3d2c355-fea3-4828-8291-77d0daa90ee5" (UID: "e3d2c355-fea3-4828-8291-77d0daa90ee5"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.879870 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e3d2c355-fea3-4828-8291-77d0daa90ee5" (UID: "e3d2c355-fea3-4828-8291-77d0daa90ee5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.917065 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.917135 5025 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.917168 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6rhh\" (UniqueName: \"kubernetes.io/projected/e3d2c355-fea3-4828-8291-77d0daa90ee5-kube-api-access-h6rhh\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.917196 5025 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/e3d2c355-fea3-4828-8291-77d0daa90ee5-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:29 crc kubenswrapper[5025]: I1004 11:08:29.917214 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3d2c355-fea3-4828-8291-77d0daa90ee5-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.333919 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" event={"ID":"e3d2c355-fea3-4828-8291-77d0daa90ee5","Type":"ContainerDied","Data":"b50bf7cd9b7e47a883d3f29d2319383764e5c840296bff6c3c66931ecead380c"} Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.333978 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b50bf7cd9b7e47a883d3f29d2319383764e5c840296bff6c3c66931ecead380c" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.334118 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-42n5d" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.488347 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t"] Oct 04 11:08:30 crc kubenswrapper[5025]: E1004 11:08:30.488998 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerName="extract-utilities" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.489070 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerName="extract-utilities" Oct 04 11:08:30 crc kubenswrapper[5025]: E1004 11:08:30.489113 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3d2c355-fea3-4828-8291-77d0daa90ee5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.489126 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3d2c355-fea3-4828-8291-77d0daa90ee5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 11:08:30 crc kubenswrapper[5025]: E1004 11:08:30.489163 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerName="extract-content" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.489176 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerName="extract-content" Oct 04 11:08:30 crc kubenswrapper[5025]: E1004 11:08:30.489243 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerName="registry-server" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.489259 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerName="registry-server" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.489682 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3d2c355-fea3-4828-8291-77d0daa90ee5" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.489727 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4f4069d-ec58-4222-b66e-8fc60955d6ba" containerName="registry-server" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.490961 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.492902 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.493077 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.495114 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.495271 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.495548 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.495786 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.502708 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t"] Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.535935 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.536154 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.536340 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.536383 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.536818 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.537123 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5t7r\" (UniqueName: \"kubernetes.io/projected/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-kube-api-access-c5t7r\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.639138 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.639208 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5t7r\" (UniqueName: \"kubernetes.io/projected/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-kube-api-access-c5t7r\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.639252 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.639283 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.639328 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.639349 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.643117 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.643481 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.643541 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.643948 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.644246 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.659815 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5t7r\" (UniqueName: \"kubernetes.io/projected/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-kube-api-access-c5t7r\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:30 crc kubenswrapper[5025]: I1004 11:08:30.814713 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:08:31 crc kubenswrapper[5025]: I1004 11:08:31.407822 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t"] Oct 04 11:08:32 crc kubenswrapper[5025]: I1004 11:08:32.356170 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" event={"ID":"b85d77e5-edce-41bf-9b7f-d1cdda6861f3","Type":"ContainerStarted","Data":"1adbbdda629bdd02a0d9220aca1188e53c72e956e0b257b936751aa08c6f7b0c"} Oct 04 11:08:32 crc kubenswrapper[5025]: I1004 11:08:32.356409 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" event={"ID":"b85d77e5-edce-41bf-9b7f-d1cdda6861f3","Type":"ContainerStarted","Data":"a6dbdc310fe41dd0183da1fb28bf67a3adceb6959b9ca61b4cf9edaca3aa9979"} Oct 04 11:08:32 crc kubenswrapper[5025]: I1004 11:08:32.380066 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" podStartSLOduration=1.74714667 podStartE2EDuration="2.380044516s" podCreationTimestamp="2025-10-04 11:08:30 +0000 UTC" firstStartedPulling="2025-10-04 11:08:31.418129345 +0000 UTC m=+2039.843096255" lastFinishedPulling="2025-10-04 11:08:32.051027221 +0000 UTC m=+2040.475994101" observedRunningTime="2025-10-04 11:08:32.376188473 +0000 UTC m=+2040.801155373" watchObservedRunningTime="2025-10-04 11:08:32.380044516 +0000 UTC m=+2040.805011406" Oct 04 11:09:20 crc kubenswrapper[5025]: I1004 11:09:20.901771 5025 generic.go:334] "Generic (PLEG): container finished" podID="b85d77e5-edce-41bf-9b7f-d1cdda6861f3" containerID="1adbbdda629bdd02a0d9220aca1188e53c72e956e0b257b936751aa08c6f7b0c" exitCode=0 Oct 04 11:09:20 crc kubenswrapper[5025]: I1004 11:09:20.901887 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" event={"ID":"b85d77e5-edce-41bf-9b7f-d1cdda6861f3","Type":"ContainerDied","Data":"1adbbdda629bdd02a0d9220aca1188e53c72e956e0b257b936751aa08c6f7b0c"} Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.368392 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.492076 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5t7r\" (UniqueName: \"kubernetes.io/projected/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-kube-api-access-c5t7r\") pod \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.492187 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-metadata-combined-ca-bundle\") pod \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.492239 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-inventory\") pod \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.492256 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-ovn-metadata-agent-neutron-config-0\") pod \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.492315 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-nova-metadata-neutron-config-0\") pod \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.492422 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-ssh-key\") pod \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\" (UID: \"b85d77e5-edce-41bf-9b7f-d1cdda6861f3\") " Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.501298 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b85d77e5-edce-41bf-9b7f-d1cdda6861f3" (UID: "b85d77e5-edce-41bf-9b7f-d1cdda6861f3"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.503465 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-kube-api-access-c5t7r" (OuterVolumeSpecName: "kube-api-access-c5t7r") pod "b85d77e5-edce-41bf-9b7f-d1cdda6861f3" (UID: "b85d77e5-edce-41bf-9b7f-d1cdda6861f3"). InnerVolumeSpecName "kube-api-access-c5t7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.520207 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "b85d77e5-edce-41bf-9b7f-d1cdda6861f3" (UID: "b85d77e5-edce-41bf-9b7f-d1cdda6861f3"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.521244 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b85d77e5-edce-41bf-9b7f-d1cdda6861f3" (UID: "b85d77e5-edce-41bf-9b7f-d1cdda6861f3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.529533 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "b85d77e5-edce-41bf-9b7f-d1cdda6861f3" (UID: "b85d77e5-edce-41bf-9b7f-d1cdda6861f3"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.538964 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-inventory" (OuterVolumeSpecName: "inventory") pod "b85d77e5-edce-41bf-9b7f-d1cdda6861f3" (UID: "b85d77e5-edce-41bf-9b7f-d1cdda6861f3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.594863 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5t7r\" (UniqueName: \"kubernetes.io/projected/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-kube-api-access-c5t7r\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.594899 5025 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.594914 5025 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.594929 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.594941 5025 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.594953 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b85d77e5-edce-41bf-9b7f-d1cdda6861f3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.929086 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" event={"ID":"b85d77e5-edce-41bf-9b7f-d1cdda6861f3","Type":"ContainerDied","Data":"a6dbdc310fe41dd0183da1fb28bf67a3adceb6959b9ca61b4cf9edaca3aa9979"} Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.929674 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6dbdc310fe41dd0183da1fb28bf67a3adceb6959b9ca61b4cf9edaca3aa9979" Oct 04 11:09:22 crc kubenswrapper[5025]: I1004 11:09:22.929806 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.031418 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987"] Oct 04 11:09:23 crc kubenswrapper[5025]: E1004 11:09:23.031832 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b85d77e5-edce-41bf-9b7f-d1cdda6861f3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.031849 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b85d77e5-edce-41bf-9b7f-d1cdda6861f3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.032069 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="b85d77e5-edce-41bf-9b7f-d1cdda6861f3" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.032686 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.035492 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.035688 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.035769 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.035713 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.035962 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.041199 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987"] Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.205931 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.206104 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvr4q\" (UniqueName: \"kubernetes.io/projected/67bb3dde-fd47-406d-8862-b365032f6ac9-kube-api-access-fvr4q\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.206226 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.206312 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.206415 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.308120 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.308236 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.308309 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvr4q\" (UniqueName: \"kubernetes.io/projected/67bb3dde-fd47-406d-8862-b365032f6ac9-kube-api-access-fvr4q\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.308391 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.308457 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.313912 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.314253 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.315800 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.319832 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.328853 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvr4q\" (UniqueName: \"kubernetes.io/projected/67bb3dde-fd47-406d-8862-b365032f6ac9-kube-api-access-fvr4q\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6k987\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.359361 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.870890 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987"] Oct 04 11:09:23 crc kubenswrapper[5025]: I1004 11:09:23.939817 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" event={"ID":"67bb3dde-fd47-406d-8862-b365032f6ac9","Type":"ContainerStarted","Data":"f5d663a2f4864a40eeb485d88ecf597c288a083ea1b2d460f29290356bf47dd9"} Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.184825 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kvzmb"] Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.188193 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.200549 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kvzmb"] Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.325872 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9sll\" (UniqueName: \"kubernetes.io/projected/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-kube-api-access-v9sll\") pod \"redhat-operators-kvzmb\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.326041 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-catalog-content\") pod \"redhat-operators-kvzmb\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.326386 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-utilities\") pod \"redhat-operators-kvzmb\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.428912 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-utilities\") pod \"redhat-operators-kvzmb\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.429048 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9sll\" (UniqueName: \"kubernetes.io/projected/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-kube-api-access-v9sll\") pod \"redhat-operators-kvzmb\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.429135 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-catalog-content\") pod \"redhat-operators-kvzmb\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.429440 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-utilities\") pod \"redhat-operators-kvzmb\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.429678 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-catalog-content\") pod \"redhat-operators-kvzmb\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.450519 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9sll\" (UniqueName: \"kubernetes.io/projected/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-kube-api-access-v9sll\") pod \"redhat-operators-kvzmb\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.528253 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.951987 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" event={"ID":"67bb3dde-fd47-406d-8862-b365032f6ac9","Type":"ContainerStarted","Data":"be1342127e9d098bd608fe27b0da6828741ee4e70971a10cdbd29acb2c275250"} Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.967435 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" podStartSLOduration=2.412147936 podStartE2EDuration="2.967417725s" podCreationTimestamp="2025-10-04 11:09:22 +0000 UTC" firstStartedPulling="2025-10-04 11:09:23.876812799 +0000 UTC m=+2092.301779679" lastFinishedPulling="2025-10-04 11:09:24.432082588 +0000 UTC m=+2092.857049468" observedRunningTime="2025-10-04 11:09:24.965371735 +0000 UTC m=+2093.390338615" watchObservedRunningTime="2025-10-04 11:09:24.967417725 +0000 UTC m=+2093.392384605" Oct 04 11:09:24 crc kubenswrapper[5025]: I1004 11:09:24.999278 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kvzmb"] Oct 04 11:09:25 crc kubenswrapper[5025]: W1004 11:09:25.000888 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d34237b_2f2a_4d7b_89c7_c8f0a8fcabd4.slice/crio-e0d2b2ea161cc0847d8ac43e250ab68b8bbfd3c1c7e082210c5f2f06b637444e WatchSource:0}: Error finding container e0d2b2ea161cc0847d8ac43e250ab68b8bbfd3c1c7e082210c5f2f06b637444e: Status 404 returned error can't find the container with id e0d2b2ea161cc0847d8ac43e250ab68b8bbfd3c1c7e082210c5f2f06b637444e Oct 04 11:09:25 crc kubenswrapper[5025]: I1004 11:09:25.972313 5025 generic.go:334] "Generic (PLEG): container finished" podID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerID="a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4" exitCode=0 Oct 04 11:09:25 crc kubenswrapper[5025]: I1004 11:09:25.972495 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvzmb" event={"ID":"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4","Type":"ContainerDied","Data":"a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4"} Oct 04 11:09:25 crc kubenswrapper[5025]: I1004 11:09:25.972715 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvzmb" event={"ID":"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4","Type":"ContainerStarted","Data":"e0d2b2ea161cc0847d8ac43e250ab68b8bbfd3c1c7e082210c5f2f06b637444e"} Oct 04 11:09:26 crc kubenswrapper[5025]: I1004 11:09:26.985442 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvzmb" event={"ID":"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4","Type":"ContainerStarted","Data":"762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f"} Oct 04 11:09:27 crc kubenswrapper[5025]: I1004 11:09:27.997728 5025 generic.go:334] "Generic (PLEG): container finished" podID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerID="762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f" exitCode=0 Oct 04 11:09:27 crc kubenswrapper[5025]: I1004 11:09:27.997796 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvzmb" event={"ID":"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4","Type":"ContainerDied","Data":"762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f"} Oct 04 11:09:29 crc kubenswrapper[5025]: I1004 11:09:29.008739 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvzmb" event={"ID":"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4","Type":"ContainerStarted","Data":"f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7"} Oct 04 11:09:29 crc kubenswrapper[5025]: I1004 11:09:29.033331 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kvzmb" podStartSLOduration=2.5839026130000002 podStartE2EDuration="5.033311486s" podCreationTimestamp="2025-10-04 11:09:24 +0000 UTC" firstStartedPulling="2025-10-04 11:09:25.975236205 +0000 UTC m=+2094.400203085" lastFinishedPulling="2025-10-04 11:09:28.424645078 +0000 UTC m=+2096.849611958" observedRunningTime="2025-10-04 11:09:29.028368102 +0000 UTC m=+2097.453334982" watchObservedRunningTime="2025-10-04 11:09:29.033311486 +0000 UTC m=+2097.458278366" Oct 04 11:09:34 crc kubenswrapper[5025]: I1004 11:09:34.528913 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:34 crc kubenswrapper[5025]: I1004 11:09:34.529858 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:34 crc kubenswrapper[5025]: I1004 11:09:34.586259 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:35 crc kubenswrapper[5025]: I1004 11:09:35.158780 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:35 crc kubenswrapper[5025]: I1004 11:09:35.219857 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kvzmb"] Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.092793 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kvzmb" podUID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerName="registry-server" containerID="cri-o://f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7" gracePeriod=2 Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.620492 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.717685 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-utilities\") pod \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.718009 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9sll\" (UniqueName: \"kubernetes.io/projected/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-kube-api-access-v9sll\") pod \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.718201 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-catalog-content\") pod \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\" (UID: \"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4\") " Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.718636 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-utilities" (OuterVolumeSpecName: "utilities") pod "9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" (UID: "9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.719159 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.722977 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-kube-api-access-v9sll" (OuterVolumeSpecName: "kube-api-access-v9sll") pod "9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" (UID: "9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4"). InnerVolumeSpecName "kube-api-access-v9sll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.809902 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" (UID: "9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.821175 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9sll\" (UniqueName: \"kubernetes.io/projected/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-kube-api-access-v9sll\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:37 crc kubenswrapper[5025]: I1004 11:09:37.821207 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.106051 5025 generic.go:334] "Generic (PLEG): container finished" podID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerID="f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7" exitCode=0 Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.106104 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvzmb" event={"ID":"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4","Type":"ContainerDied","Data":"f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7"} Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.106140 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kvzmb" event={"ID":"9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4","Type":"ContainerDied","Data":"e0d2b2ea161cc0847d8ac43e250ab68b8bbfd3c1c7e082210c5f2f06b637444e"} Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.106136 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kvzmb" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.106157 5025 scope.go:117] "RemoveContainer" containerID="f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.125082 5025 scope.go:117] "RemoveContainer" containerID="762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.156397 5025 scope.go:117] "RemoveContainer" containerID="a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.164199 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kvzmb"] Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.172357 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kvzmb"] Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.198371 5025 scope.go:117] "RemoveContainer" containerID="f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7" Oct 04 11:09:38 crc kubenswrapper[5025]: E1004 11:09:38.198946 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7\": container with ID starting with f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7 not found: ID does not exist" containerID="f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.198998 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7"} err="failed to get container status \"f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7\": rpc error: code = NotFound desc = could not find container \"f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7\": container with ID starting with f3d4696cc0ed58f04c49c484d73b44f1d6e21ba7c7e20f08a47943d0362d4aa7 not found: ID does not exist" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.199049 5025 scope.go:117] "RemoveContainer" containerID="762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f" Oct 04 11:09:38 crc kubenswrapper[5025]: E1004 11:09:38.199477 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f\": container with ID starting with 762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f not found: ID does not exist" containerID="762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.199517 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f"} err="failed to get container status \"762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f\": rpc error: code = NotFound desc = could not find container \"762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f\": container with ID starting with 762e50a60268f61df57a19a986a7ca84ac8284cfab01865ea1dc74f05abece5f not found: ID does not exist" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.199544 5025 scope.go:117] "RemoveContainer" containerID="a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4" Oct 04 11:09:38 crc kubenswrapper[5025]: E1004 11:09:38.199845 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4\": container with ID starting with a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4 not found: ID does not exist" containerID="a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.199894 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4"} err="failed to get container status \"a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4\": rpc error: code = NotFound desc = could not find container \"a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4\": container with ID starting with a6f947a0c026ab512c8626a1e0106117e7e8a0a0cc769cad1593bcb9afac1eb4 not found: ID does not exist" Oct 04 11:09:38 crc kubenswrapper[5025]: I1004 11:09:38.425971 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" path="/var/lib/kubelet/pods/9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4/volumes" Oct 04 11:09:44 crc kubenswrapper[5025]: I1004 11:09:44.713472 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:09:44 crc kubenswrapper[5025]: I1004 11:09:44.714182 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:10:14 crc kubenswrapper[5025]: I1004 11:10:14.713785 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:10:14 crc kubenswrapper[5025]: I1004 11:10:14.714302 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.359053 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nd75c"] Oct 04 11:10:28 crc kubenswrapper[5025]: E1004 11:10:28.360256 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerName="registry-server" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.360278 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerName="registry-server" Oct 04 11:10:28 crc kubenswrapper[5025]: E1004 11:10:28.360318 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerName="extract-content" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.360334 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerName="extract-content" Oct 04 11:10:28 crc kubenswrapper[5025]: E1004 11:10:28.360359 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerName="extract-utilities" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.360373 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerName="extract-utilities" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.360696 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d34237b-2f2a-4d7b-89c7-c8f0a8fcabd4" containerName="registry-server" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.363346 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.370515 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nd75c"] Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.461967 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-utilities\") pod \"certified-operators-nd75c\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.462050 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sfvg\" (UniqueName: \"kubernetes.io/projected/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-kube-api-access-7sfvg\") pod \"certified-operators-nd75c\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.462127 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-catalog-content\") pod \"certified-operators-nd75c\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.563949 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-catalog-content\") pod \"certified-operators-nd75c\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.564126 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-utilities\") pod \"certified-operators-nd75c\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.564163 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sfvg\" (UniqueName: \"kubernetes.io/projected/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-kube-api-access-7sfvg\") pod \"certified-operators-nd75c\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.564615 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-utilities\") pod \"certified-operators-nd75c\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.565754 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-catalog-content\") pod \"certified-operators-nd75c\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.583750 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sfvg\" (UniqueName: \"kubernetes.io/projected/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-kube-api-access-7sfvg\") pod \"certified-operators-nd75c\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:28 crc kubenswrapper[5025]: I1004 11:10:28.692535 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:29 crc kubenswrapper[5025]: I1004 11:10:29.193176 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nd75c"] Oct 04 11:10:29 crc kubenswrapper[5025]: I1004 11:10:29.653280 5025 generic.go:334] "Generic (PLEG): container finished" podID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerID="2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f" exitCode=0 Oct 04 11:10:29 crc kubenswrapper[5025]: I1004 11:10:29.653397 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd75c" event={"ID":"95fa1bf6-e3fd-44fd-b254-c5acd5b18596","Type":"ContainerDied","Data":"2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f"} Oct 04 11:10:29 crc kubenswrapper[5025]: I1004 11:10:29.653675 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd75c" event={"ID":"95fa1bf6-e3fd-44fd-b254-c5acd5b18596","Type":"ContainerStarted","Data":"7af6767a9ec0b5fe684c4569757b8f838f650febc2b1519d849dc547582fe4c3"} Oct 04 11:10:31 crc kubenswrapper[5025]: I1004 11:10:31.679720 5025 generic.go:334] "Generic (PLEG): container finished" podID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerID="84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464" exitCode=0 Oct 04 11:10:31 crc kubenswrapper[5025]: I1004 11:10:31.679802 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd75c" event={"ID":"95fa1bf6-e3fd-44fd-b254-c5acd5b18596","Type":"ContainerDied","Data":"84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464"} Oct 04 11:10:33 crc kubenswrapper[5025]: I1004 11:10:33.701351 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd75c" event={"ID":"95fa1bf6-e3fd-44fd-b254-c5acd5b18596","Type":"ContainerStarted","Data":"d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35"} Oct 04 11:10:33 crc kubenswrapper[5025]: I1004 11:10:33.724087 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nd75c" podStartSLOduration=2.930937283 podStartE2EDuration="5.724061621s" podCreationTimestamp="2025-10-04 11:10:28 +0000 UTC" firstStartedPulling="2025-10-04 11:10:29.655560845 +0000 UTC m=+2158.080527725" lastFinishedPulling="2025-10-04 11:10:32.448685183 +0000 UTC m=+2160.873652063" observedRunningTime="2025-10-04 11:10:33.717584679 +0000 UTC m=+2162.142551559" watchObservedRunningTime="2025-10-04 11:10:33.724061621 +0000 UTC m=+2162.149028521" Oct 04 11:10:38 crc kubenswrapper[5025]: I1004 11:10:38.692759 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:38 crc kubenswrapper[5025]: I1004 11:10:38.693609 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:38 crc kubenswrapper[5025]: I1004 11:10:38.744420 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:38 crc kubenswrapper[5025]: I1004 11:10:38.802163 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:38 crc kubenswrapper[5025]: I1004 11:10:38.986568 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nd75c"] Oct 04 11:10:40 crc kubenswrapper[5025]: I1004 11:10:40.775380 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nd75c" podUID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerName="registry-server" containerID="cri-o://d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35" gracePeriod=2 Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.499354 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.654098 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-utilities\") pod \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.654144 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sfvg\" (UniqueName: \"kubernetes.io/projected/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-kube-api-access-7sfvg\") pod \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.654209 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-catalog-content\") pod \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\" (UID: \"95fa1bf6-e3fd-44fd-b254-c5acd5b18596\") " Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.655757 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-utilities" (OuterVolumeSpecName: "utilities") pod "95fa1bf6-e3fd-44fd-b254-c5acd5b18596" (UID: "95fa1bf6-e3fd-44fd-b254-c5acd5b18596"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.659524 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.665364 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-kube-api-access-7sfvg" (OuterVolumeSpecName: "kube-api-access-7sfvg") pod "95fa1bf6-e3fd-44fd-b254-c5acd5b18596" (UID: "95fa1bf6-e3fd-44fd-b254-c5acd5b18596"). InnerVolumeSpecName "kube-api-access-7sfvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.706391 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95fa1bf6-e3fd-44fd-b254-c5acd5b18596" (UID: "95fa1bf6-e3fd-44fd-b254-c5acd5b18596"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.761981 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sfvg\" (UniqueName: \"kubernetes.io/projected/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-kube-api-access-7sfvg\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.762028 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95fa1bf6-e3fd-44fd-b254-c5acd5b18596-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.787483 5025 generic.go:334] "Generic (PLEG): container finished" podID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerID="d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35" exitCode=0 Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.787526 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd75c" event={"ID":"95fa1bf6-e3fd-44fd-b254-c5acd5b18596","Type":"ContainerDied","Data":"d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35"} Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.787560 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd75c" event={"ID":"95fa1bf6-e3fd-44fd-b254-c5acd5b18596","Type":"ContainerDied","Data":"7af6767a9ec0b5fe684c4569757b8f838f650febc2b1519d849dc547582fe4c3"} Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.787555 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd75c" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.787673 5025 scope.go:117] "RemoveContainer" containerID="d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.820633 5025 scope.go:117] "RemoveContainer" containerID="84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.835736 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nd75c"] Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.847029 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nd75c"] Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.853498 5025 scope.go:117] "RemoveContainer" containerID="2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.902915 5025 scope.go:117] "RemoveContainer" containerID="d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35" Oct 04 11:10:41 crc kubenswrapper[5025]: E1004 11:10:41.903493 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35\": container with ID starting with d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35 not found: ID does not exist" containerID="d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.903530 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35"} err="failed to get container status \"d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35\": rpc error: code = NotFound desc = could not find container \"d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35\": container with ID starting with d4e9c302207711a42c29fc3471c98cdf5db6d0842848a776a7728a96fc73be35 not found: ID does not exist" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.903557 5025 scope.go:117] "RemoveContainer" containerID="84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464" Oct 04 11:10:41 crc kubenswrapper[5025]: E1004 11:10:41.904194 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464\": container with ID starting with 84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464 not found: ID does not exist" containerID="84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.904214 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464"} err="failed to get container status \"84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464\": rpc error: code = NotFound desc = could not find container \"84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464\": container with ID starting with 84ad222f7271d7c46e10981505b8b6f3aaf70048faf65393eb0045778ea05464 not found: ID does not exist" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.904228 5025 scope.go:117] "RemoveContainer" containerID="2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f" Oct 04 11:10:41 crc kubenswrapper[5025]: E1004 11:10:41.904641 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f\": container with ID starting with 2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f not found: ID does not exist" containerID="2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f" Oct 04 11:10:41 crc kubenswrapper[5025]: I1004 11:10:41.904659 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f"} err="failed to get container status \"2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f\": rpc error: code = NotFound desc = could not find container \"2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f\": container with ID starting with 2f46235e560a57ed4e2ba044ec9ff5094183313f4f7d78a44cfbc6e3956ee55f not found: ID does not exist" Oct 04 11:10:42 crc kubenswrapper[5025]: I1004 11:10:42.426535 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" path="/var/lib/kubelet/pods/95fa1bf6-e3fd-44fd-b254-c5acd5b18596/volumes" Oct 04 11:10:44 crc kubenswrapper[5025]: I1004 11:10:44.713301 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:10:44 crc kubenswrapper[5025]: I1004 11:10:44.715684 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:10:44 crc kubenswrapper[5025]: I1004 11:10:44.715890 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 11:10:44 crc kubenswrapper[5025]: I1004 11:10:44.716875 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7f9c45b9be91741a2b640d8ac9c01d5e4ddf8c93a7a041740f1c11213ec28929"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:10:44 crc kubenswrapper[5025]: I1004 11:10:44.717226 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://7f9c45b9be91741a2b640d8ac9c01d5e4ddf8c93a7a041740f1c11213ec28929" gracePeriod=600 Oct 04 11:10:45 crc kubenswrapper[5025]: I1004 11:10:45.839296 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="7f9c45b9be91741a2b640d8ac9c01d5e4ddf8c93a7a041740f1c11213ec28929" exitCode=0 Oct 04 11:10:45 crc kubenswrapper[5025]: I1004 11:10:45.839395 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"7f9c45b9be91741a2b640d8ac9c01d5e4ddf8c93a7a041740f1c11213ec28929"} Oct 04 11:10:45 crc kubenswrapper[5025]: I1004 11:10:45.839998 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa"} Oct 04 11:10:45 crc kubenswrapper[5025]: I1004 11:10:45.840048 5025 scope.go:117] "RemoveContainer" containerID="61f81be8f72c4409045cf707b5bb8c86ea2b6c1fbcdea051e625f8402ef82634" Oct 04 11:11:50 crc kubenswrapper[5025]: I1004 11:11:50.867936 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xh95x"] Oct 04 11:11:50 crc kubenswrapper[5025]: E1004 11:11:50.869044 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerName="registry-server" Oct 04 11:11:50 crc kubenswrapper[5025]: I1004 11:11:50.869065 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerName="registry-server" Oct 04 11:11:50 crc kubenswrapper[5025]: E1004 11:11:50.869079 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerName="extract-utilities" Oct 04 11:11:50 crc kubenswrapper[5025]: I1004 11:11:50.869088 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerName="extract-utilities" Oct 04 11:11:50 crc kubenswrapper[5025]: E1004 11:11:50.869115 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerName="extract-content" Oct 04 11:11:50 crc kubenswrapper[5025]: I1004 11:11:50.869125 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerName="extract-content" Oct 04 11:11:50 crc kubenswrapper[5025]: I1004 11:11:50.869420 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="95fa1bf6-e3fd-44fd-b254-c5acd5b18596" containerName="registry-server" Oct 04 11:11:50 crc kubenswrapper[5025]: I1004 11:11:50.874402 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:50 crc kubenswrapper[5025]: I1004 11:11:50.881434 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xh95x"] Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.015129 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-utilities\") pod \"community-operators-xh95x\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.015271 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-catalog-content\") pod \"community-operators-xh95x\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.015321 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdnw6\" (UniqueName: \"kubernetes.io/projected/3c3b64df-de62-4a6c-94af-c73b7e702e4a-kube-api-access-hdnw6\") pod \"community-operators-xh95x\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.116400 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-catalog-content\") pod \"community-operators-xh95x\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.116487 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdnw6\" (UniqueName: \"kubernetes.io/projected/3c3b64df-de62-4a6c-94af-c73b7e702e4a-kube-api-access-hdnw6\") pod \"community-operators-xh95x\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.116557 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-utilities\") pod \"community-operators-xh95x\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.116923 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-catalog-content\") pod \"community-operators-xh95x\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.117035 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-utilities\") pod \"community-operators-xh95x\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.137337 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdnw6\" (UniqueName: \"kubernetes.io/projected/3c3b64df-de62-4a6c-94af-c73b7e702e4a-kube-api-access-hdnw6\") pod \"community-operators-xh95x\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.210316 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:11:51 crc kubenswrapper[5025]: I1004 11:11:51.687888 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xh95x"] Oct 04 11:11:52 crc kubenswrapper[5025]: I1004 11:11:52.497704 5025 generic.go:334] "Generic (PLEG): container finished" podID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerID="144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b" exitCode=0 Oct 04 11:11:52 crc kubenswrapper[5025]: I1004 11:11:52.497828 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xh95x" event={"ID":"3c3b64df-de62-4a6c-94af-c73b7e702e4a","Type":"ContainerDied","Data":"144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b"} Oct 04 11:11:52 crc kubenswrapper[5025]: I1004 11:11:52.500302 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xh95x" event={"ID":"3c3b64df-de62-4a6c-94af-c73b7e702e4a","Type":"ContainerStarted","Data":"2eb966e248066981e9a0c1d0c1120167c87ddaa446ea7a36028763bd725ac715"} Oct 04 11:11:53 crc kubenswrapper[5025]: I1004 11:11:53.514818 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xh95x" event={"ID":"3c3b64df-de62-4a6c-94af-c73b7e702e4a","Type":"ContainerStarted","Data":"b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381"} Oct 04 11:11:54 crc kubenswrapper[5025]: I1004 11:11:54.524869 5025 generic.go:334] "Generic (PLEG): container finished" podID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerID="b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381" exitCode=0 Oct 04 11:11:54 crc kubenswrapper[5025]: I1004 11:11:54.524958 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xh95x" event={"ID":"3c3b64df-de62-4a6c-94af-c73b7e702e4a","Type":"ContainerDied","Data":"b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381"} Oct 04 11:11:55 crc kubenswrapper[5025]: I1004 11:11:55.535922 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xh95x" event={"ID":"3c3b64df-de62-4a6c-94af-c73b7e702e4a","Type":"ContainerStarted","Data":"f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5"} Oct 04 11:11:55 crc kubenswrapper[5025]: I1004 11:11:55.567244 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xh95x" podStartSLOduration=2.857532578 podStartE2EDuration="5.567224998s" podCreationTimestamp="2025-10-04 11:11:50 +0000 UTC" firstStartedPulling="2025-10-04 11:11:52.499930805 +0000 UTC m=+2240.924897685" lastFinishedPulling="2025-10-04 11:11:55.209623225 +0000 UTC m=+2243.634590105" observedRunningTime="2025-10-04 11:11:55.56482155 +0000 UTC m=+2243.989788440" watchObservedRunningTime="2025-10-04 11:11:55.567224998 +0000 UTC m=+2243.992191888" Oct 04 11:12:01 crc kubenswrapper[5025]: I1004 11:12:01.210919 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:12:01 crc kubenswrapper[5025]: I1004 11:12:01.211315 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:12:01 crc kubenswrapper[5025]: I1004 11:12:01.265487 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:12:01 crc kubenswrapper[5025]: I1004 11:12:01.649484 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:12:02 crc kubenswrapper[5025]: I1004 11:12:02.659547 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xh95x"] Oct 04 11:12:03 crc kubenswrapper[5025]: I1004 11:12:03.614813 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xh95x" podUID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerName="registry-server" containerID="cri-o://f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5" gracePeriod=2 Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.058172 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.077843 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdnw6\" (UniqueName: \"kubernetes.io/projected/3c3b64df-de62-4a6c-94af-c73b7e702e4a-kube-api-access-hdnw6\") pod \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.078043 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-utilities\") pod \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.078118 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-catalog-content\") pod \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\" (UID: \"3c3b64df-de62-4a6c-94af-c73b7e702e4a\") " Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.078914 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-utilities" (OuterVolumeSpecName: "utilities") pod "3c3b64df-de62-4a6c-94af-c73b7e702e4a" (UID: "3c3b64df-de62-4a6c-94af-c73b7e702e4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.084620 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c3b64df-de62-4a6c-94af-c73b7e702e4a-kube-api-access-hdnw6" (OuterVolumeSpecName: "kube-api-access-hdnw6") pod "3c3b64df-de62-4a6c-94af-c73b7e702e4a" (UID: "3c3b64df-de62-4a6c-94af-c73b7e702e4a"). InnerVolumeSpecName "kube-api-access-hdnw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.146074 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c3b64df-de62-4a6c-94af-c73b7e702e4a" (UID: "3c3b64df-de62-4a6c-94af-c73b7e702e4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.180318 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdnw6\" (UniqueName: \"kubernetes.io/projected/3c3b64df-de62-4a6c-94af-c73b7e702e4a-kube-api-access-hdnw6\") on node \"crc\" DevicePath \"\"" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.180362 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.180398 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3b64df-de62-4a6c-94af-c73b7e702e4a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.625763 5025 generic.go:334] "Generic (PLEG): container finished" podID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerID="f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5" exitCode=0 Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.625824 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xh95x" event={"ID":"3c3b64df-de62-4a6c-94af-c73b7e702e4a","Type":"ContainerDied","Data":"f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5"} Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.625857 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xh95x" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.625885 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xh95x" event={"ID":"3c3b64df-de62-4a6c-94af-c73b7e702e4a","Type":"ContainerDied","Data":"2eb966e248066981e9a0c1d0c1120167c87ddaa446ea7a36028763bd725ac715"} Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.625911 5025 scope.go:117] "RemoveContainer" containerID="f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.653315 5025 scope.go:117] "RemoveContainer" containerID="b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.657489 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xh95x"] Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.665521 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xh95x"] Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.682364 5025 scope.go:117] "RemoveContainer" containerID="144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.731419 5025 scope.go:117] "RemoveContainer" containerID="f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5" Oct 04 11:12:04 crc kubenswrapper[5025]: E1004 11:12:04.732040 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5\": container with ID starting with f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5 not found: ID does not exist" containerID="f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.732070 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5"} err="failed to get container status \"f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5\": rpc error: code = NotFound desc = could not find container \"f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5\": container with ID starting with f252c13c2852b2bcc51d0756eb5b779c8dedf6aac10080bd2ff63eb941ffd8f5 not found: ID does not exist" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.732091 5025 scope.go:117] "RemoveContainer" containerID="b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381" Oct 04 11:12:04 crc kubenswrapper[5025]: E1004 11:12:04.732504 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381\": container with ID starting with b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381 not found: ID does not exist" containerID="b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.732552 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381"} err="failed to get container status \"b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381\": rpc error: code = NotFound desc = could not find container \"b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381\": container with ID starting with b4315af7e510aa8511b1f1e24019aa267ca5b5f8efe283e45bb8905c13d22381 not found: ID does not exist" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.732588 5025 scope.go:117] "RemoveContainer" containerID="144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b" Oct 04 11:12:04 crc kubenswrapper[5025]: E1004 11:12:04.733115 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b\": container with ID starting with 144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b not found: ID does not exist" containerID="144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b" Oct 04 11:12:04 crc kubenswrapper[5025]: I1004 11:12:04.733160 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b"} err="failed to get container status \"144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b\": rpc error: code = NotFound desc = could not find container \"144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b\": container with ID starting with 144e6bad3ac3f1c2e2626d508f603cbffe0088a8c3def18815f732d42cf9a76b not found: ID does not exist" Oct 04 11:12:06 crc kubenswrapper[5025]: I1004 11:12:06.452722 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" path="/var/lib/kubelet/pods/3c3b64df-de62-4a6c-94af-c73b7e702e4a/volumes" Oct 04 11:13:14 crc kubenswrapper[5025]: I1004 11:13:14.713909 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:13:14 crc kubenswrapper[5025]: I1004 11:13:14.714730 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:13:34 crc kubenswrapper[5025]: I1004 11:13:34.513691 5025 generic.go:334] "Generic (PLEG): container finished" podID="67bb3dde-fd47-406d-8862-b365032f6ac9" containerID="be1342127e9d098bd608fe27b0da6828741ee4e70971a10cdbd29acb2c275250" exitCode=0 Oct 04 11:13:34 crc kubenswrapper[5025]: I1004 11:13:34.513798 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" event={"ID":"67bb3dde-fd47-406d-8862-b365032f6ac9","Type":"ContainerDied","Data":"be1342127e9d098bd608fe27b0da6828741ee4e70971a10cdbd29acb2c275250"} Oct 04 11:13:35 crc kubenswrapper[5025]: I1004 11:13:35.961810 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.021903 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvr4q\" (UniqueName: \"kubernetes.io/projected/67bb3dde-fd47-406d-8862-b365032f6ac9-kube-api-access-fvr4q\") pod \"67bb3dde-fd47-406d-8862-b365032f6ac9\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.022109 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-combined-ca-bundle\") pod \"67bb3dde-fd47-406d-8862-b365032f6ac9\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.022154 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-ssh-key\") pod \"67bb3dde-fd47-406d-8862-b365032f6ac9\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.022254 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-inventory\") pod \"67bb3dde-fd47-406d-8862-b365032f6ac9\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.022398 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-secret-0\") pod \"67bb3dde-fd47-406d-8862-b365032f6ac9\" (UID: \"67bb3dde-fd47-406d-8862-b365032f6ac9\") " Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.027613 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "67bb3dde-fd47-406d-8862-b365032f6ac9" (UID: "67bb3dde-fd47-406d-8862-b365032f6ac9"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.032231 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67bb3dde-fd47-406d-8862-b365032f6ac9-kube-api-access-fvr4q" (OuterVolumeSpecName: "kube-api-access-fvr4q") pod "67bb3dde-fd47-406d-8862-b365032f6ac9" (UID: "67bb3dde-fd47-406d-8862-b365032f6ac9"). InnerVolumeSpecName "kube-api-access-fvr4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.049724 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "67bb3dde-fd47-406d-8862-b365032f6ac9" (UID: "67bb3dde-fd47-406d-8862-b365032f6ac9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.050141 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "67bb3dde-fd47-406d-8862-b365032f6ac9" (UID: "67bb3dde-fd47-406d-8862-b365032f6ac9"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.057902 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-inventory" (OuterVolumeSpecName: "inventory") pod "67bb3dde-fd47-406d-8862-b365032f6ac9" (UID: "67bb3dde-fd47-406d-8862-b365032f6ac9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.124317 5025 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.124350 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.124364 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.124377 5025 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/67bb3dde-fd47-406d-8862-b365032f6ac9-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.124389 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvr4q\" (UniqueName: \"kubernetes.io/projected/67bb3dde-fd47-406d-8862-b365032f6ac9-kube-api-access-fvr4q\") on node \"crc\" DevicePath \"\"" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.542525 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" event={"ID":"67bb3dde-fd47-406d-8862-b365032f6ac9","Type":"ContainerDied","Data":"f5d663a2f4864a40eeb485d88ecf597c288a083ea1b2d460f29290356bf47dd9"} Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.542574 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5d663a2f4864a40eeb485d88ecf597c288a083ea1b2d460f29290356bf47dd9" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.542581 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6k987" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.642778 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97"] Oct 04 11:13:36 crc kubenswrapper[5025]: E1004 11:13:36.643494 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerName="extract-content" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.643511 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerName="extract-content" Oct 04 11:13:36 crc kubenswrapper[5025]: E1004 11:13:36.643550 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerName="extract-utilities" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.643559 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerName="extract-utilities" Oct 04 11:13:36 crc kubenswrapper[5025]: E1004 11:13:36.643566 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67bb3dde-fd47-406d-8862-b365032f6ac9" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.643573 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="67bb3dde-fd47-406d-8862-b365032f6ac9" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 11:13:36 crc kubenswrapper[5025]: E1004 11:13:36.643596 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerName="registry-server" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.643602 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerName="registry-server" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.644008 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="67bb3dde-fd47-406d-8862-b365032f6ac9" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.646984 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3b64df-de62-4a6c-94af-c73b7e702e4a" containerName="registry-server" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.647941 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.655331 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.655331 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.655846 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.655937 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.655705 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.655790 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.655889 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.681321 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97"] Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.735552 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp677\" (UniqueName: \"kubernetes.io/projected/d807eca9-05be-475e-b168-2ab47ab555c1-kube-api-access-dp677\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.735618 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d807eca9-05be-475e-b168-2ab47ab555c1-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.735658 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.735693 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.735744 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.735768 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.735796 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.735838 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.735878 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.838124 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp677\" (UniqueName: \"kubernetes.io/projected/d807eca9-05be-475e-b168-2ab47ab555c1-kube-api-access-dp677\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.838181 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d807eca9-05be-475e-b168-2ab47ab555c1-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.838214 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.838240 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.838264 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.838283 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.838305 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.838338 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.838373 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.840915 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d807eca9-05be-475e-b168-2ab47ab555c1-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.842386 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.842446 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.842735 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.843862 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.844102 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.844457 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.846911 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.858692 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp677\" (UniqueName: \"kubernetes.io/projected/d807eca9-05be-475e-b168-2ab47ab555c1-kube-api-access-dp677\") pod \"nova-edpm-deployment-openstack-edpm-ipam-tqj97\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:36 crc kubenswrapper[5025]: I1004 11:13:36.981153 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:13:37 crc kubenswrapper[5025]: I1004 11:13:37.506657 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97"] Oct 04 11:13:37 crc kubenswrapper[5025]: I1004 11:13:37.511803 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:13:37 crc kubenswrapper[5025]: I1004 11:13:37.560134 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" event={"ID":"d807eca9-05be-475e-b168-2ab47ab555c1","Type":"ContainerStarted","Data":"9a296e0ddfc6ceda2173cb11ad9f88b42eeacf839235c1398e13d98032b4d231"} Oct 04 11:13:38 crc kubenswrapper[5025]: I1004 11:13:38.571733 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" event={"ID":"d807eca9-05be-475e-b168-2ab47ab555c1","Type":"ContainerStarted","Data":"4ecac9300bb6a7d9a0eb131034b8f54be13474f7775151be661f12664adc8a06"} Oct 04 11:13:38 crc kubenswrapper[5025]: I1004 11:13:38.597972 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" podStartSLOduration=2.020560296 podStartE2EDuration="2.597943513s" podCreationTimestamp="2025-10-04 11:13:36 +0000 UTC" firstStartedPulling="2025-10-04 11:13:37.511553352 +0000 UTC m=+2345.936520242" lastFinishedPulling="2025-10-04 11:13:38.088936579 +0000 UTC m=+2346.513903459" observedRunningTime="2025-10-04 11:13:38.595262247 +0000 UTC m=+2347.020229167" watchObservedRunningTime="2025-10-04 11:13:38.597943513 +0000 UTC m=+2347.022910433" Oct 04 11:13:44 crc kubenswrapper[5025]: I1004 11:13:44.714963 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:13:44 crc kubenswrapper[5025]: I1004 11:13:44.715916 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:14:14 crc kubenswrapper[5025]: I1004 11:14:14.714041 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:14:14 crc kubenswrapper[5025]: I1004 11:14:14.714620 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:14:14 crc kubenswrapper[5025]: I1004 11:14:14.714664 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 11:14:14 crc kubenswrapper[5025]: I1004 11:14:14.715136 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:14:14 crc kubenswrapper[5025]: I1004 11:14:14.715191 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" gracePeriod=600 Oct 04 11:14:14 crc kubenswrapper[5025]: E1004 11:14:14.869557 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:14:14 crc kubenswrapper[5025]: I1004 11:14:14.937616 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" exitCode=0 Oct 04 11:14:14 crc kubenswrapper[5025]: I1004 11:14:14.937655 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa"} Oct 04 11:14:14 crc kubenswrapper[5025]: I1004 11:14:14.937689 5025 scope.go:117] "RemoveContainer" containerID="7f9c45b9be91741a2b640d8ac9c01d5e4ddf8c93a7a041740f1c11213ec28929" Oct 04 11:14:14 crc kubenswrapper[5025]: I1004 11:14:14.938413 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:14:14 crc kubenswrapper[5025]: E1004 11:14:14.938761 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:14:30 crc kubenswrapper[5025]: I1004 11:14:30.411516 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:14:30 crc kubenswrapper[5025]: E1004 11:14:30.412256 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:14:41 crc kubenswrapper[5025]: I1004 11:14:41.411381 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:14:41 crc kubenswrapper[5025]: E1004 11:14:41.412062 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:14:55 crc kubenswrapper[5025]: I1004 11:14:55.412322 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:14:55 crc kubenswrapper[5025]: E1004 11:14:55.413241 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.165666 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l"] Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.167955 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.181525 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l"] Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.216082 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.216305 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.323480 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzmt7\" (UniqueName: \"kubernetes.io/projected/9332a111-c0da-4fc5-b9ee-462ff503750e-kube-api-access-mzmt7\") pod \"collect-profiles-29326275-5nl2l\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.323559 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9332a111-c0da-4fc5-b9ee-462ff503750e-secret-volume\") pod \"collect-profiles-29326275-5nl2l\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.323754 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9332a111-c0da-4fc5-b9ee-462ff503750e-config-volume\") pod \"collect-profiles-29326275-5nl2l\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.425832 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9332a111-c0da-4fc5-b9ee-462ff503750e-config-volume\") pod \"collect-profiles-29326275-5nl2l\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.425948 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzmt7\" (UniqueName: \"kubernetes.io/projected/9332a111-c0da-4fc5-b9ee-462ff503750e-kube-api-access-mzmt7\") pod \"collect-profiles-29326275-5nl2l\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.425986 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9332a111-c0da-4fc5-b9ee-462ff503750e-secret-volume\") pod \"collect-profiles-29326275-5nl2l\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.426838 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9332a111-c0da-4fc5-b9ee-462ff503750e-config-volume\") pod \"collect-profiles-29326275-5nl2l\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.432792 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9332a111-c0da-4fc5-b9ee-462ff503750e-secret-volume\") pod \"collect-profiles-29326275-5nl2l\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.441065 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzmt7\" (UniqueName: \"kubernetes.io/projected/9332a111-c0da-4fc5-b9ee-462ff503750e-kube-api-access-mzmt7\") pod \"collect-profiles-29326275-5nl2l\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.544528 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:00 crc kubenswrapper[5025]: I1004 11:15:00.993506 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l"] Oct 04 11:15:01 crc kubenswrapper[5025]: I1004 11:15:01.378368 5025 generic.go:334] "Generic (PLEG): container finished" podID="9332a111-c0da-4fc5-b9ee-462ff503750e" containerID="090794c0514589f0a677454426550fc078c6aa8d255e90a8e5ba0f6d617a0627" exitCode=0 Oct 04 11:15:01 crc kubenswrapper[5025]: I1004 11:15:01.378474 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" event={"ID":"9332a111-c0da-4fc5-b9ee-462ff503750e","Type":"ContainerDied","Data":"090794c0514589f0a677454426550fc078c6aa8d255e90a8e5ba0f6d617a0627"} Oct 04 11:15:01 crc kubenswrapper[5025]: I1004 11:15:01.378634 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" event={"ID":"9332a111-c0da-4fc5-b9ee-462ff503750e","Type":"ContainerStarted","Data":"f5a0ee399395d698bfcbe00566c245d53e4947f010e6c610d901fec5c1c249d2"} Oct 04 11:15:02 crc kubenswrapper[5025]: I1004 11:15:02.830577 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:02 crc kubenswrapper[5025]: I1004 11:15:02.974307 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzmt7\" (UniqueName: \"kubernetes.io/projected/9332a111-c0da-4fc5-b9ee-462ff503750e-kube-api-access-mzmt7\") pod \"9332a111-c0da-4fc5-b9ee-462ff503750e\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " Oct 04 11:15:02 crc kubenswrapper[5025]: I1004 11:15:02.974359 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9332a111-c0da-4fc5-b9ee-462ff503750e-config-volume\") pod \"9332a111-c0da-4fc5-b9ee-462ff503750e\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " Oct 04 11:15:02 crc kubenswrapper[5025]: I1004 11:15:02.974532 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9332a111-c0da-4fc5-b9ee-462ff503750e-secret-volume\") pod \"9332a111-c0da-4fc5-b9ee-462ff503750e\" (UID: \"9332a111-c0da-4fc5-b9ee-462ff503750e\") " Oct 04 11:15:02 crc kubenswrapper[5025]: I1004 11:15:02.975334 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9332a111-c0da-4fc5-b9ee-462ff503750e-config-volume" (OuterVolumeSpecName: "config-volume") pod "9332a111-c0da-4fc5-b9ee-462ff503750e" (UID: "9332a111-c0da-4fc5-b9ee-462ff503750e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:15:02 crc kubenswrapper[5025]: I1004 11:15:02.979925 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9332a111-c0da-4fc5-b9ee-462ff503750e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9332a111-c0da-4fc5-b9ee-462ff503750e" (UID: "9332a111-c0da-4fc5-b9ee-462ff503750e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:15:02 crc kubenswrapper[5025]: I1004 11:15:02.981181 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9332a111-c0da-4fc5-b9ee-462ff503750e-kube-api-access-mzmt7" (OuterVolumeSpecName: "kube-api-access-mzmt7") pod "9332a111-c0da-4fc5-b9ee-462ff503750e" (UID: "9332a111-c0da-4fc5-b9ee-462ff503750e"). InnerVolumeSpecName "kube-api-access-mzmt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:15:03 crc kubenswrapper[5025]: I1004 11:15:03.076041 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzmt7\" (UniqueName: \"kubernetes.io/projected/9332a111-c0da-4fc5-b9ee-462ff503750e-kube-api-access-mzmt7\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:03 crc kubenswrapper[5025]: I1004 11:15:03.076076 5025 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9332a111-c0da-4fc5-b9ee-462ff503750e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:03 crc kubenswrapper[5025]: I1004 11:15:03.076086 5025 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9332a111-c0da-4fc5-b9ee-462ff503750e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:15:03 crc kubenswrapper[5025]: I1004 11:15:03.403225 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" event={"ID":"9332a111-c0da-4fc5-b9ee-462ff503750e","Type":"ContainerDied","Data":"f5a0ee399395d698bfcbe00566c245d53e4947f010e6c610d901fec5c1c249d2"} Oct 04 11:15:03 crc kubenswrapper[5025]: I1004 11:15:03.403284 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326275-5nl2l" Oct 04 11:15:03 crc kubenswrapper[5025]: I1004 11:15:03.403303 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5a0ee399395d698bfcbe00566c245d53e4947f010e6c610d901fec5c1c249d2" Oct 04 11:15:03 crc kubenswrapper[5025]: I1004 11:15:03.924031 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr"] Oct 04 11:15:03 crc kubenswrapper[5025]: I1004 11:15:03.935181 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326230-67kmr"] Oct 04 11:15:04 crc kubenswrapper[5025]: I1004 11:15:04.427774 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd17fb73-1ab6-46da-9b53-34ab62c9e0fa" path="/var/lib/kubelet/pods/cd17fb73-1ab6-46da-9b53-34ab62c9e0fa/volumes" Oct 04 11:15:07 crc kubenswrapper[5025]: I1004 11:15:07.411540 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:15:07 crc kubenswrapper[5025]: E1004 11:15:07.412114 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:15:18 crc kubenswrapper[5025]: I1004 11:15:18.412110 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:15:18 crc kubenswrapper[5025]: E1004 11:15:18.413065 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:15:33 crc kubenswrapper[5025]: I1004 11:15:33.411468 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:15:33 crc kubenswrapper[5025]: E1004 11:15:33.412506 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:15:47 crc kubenswrapper[5025]: I1004 11:15:47.411871 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:15:47 crc kubenswrapper[5025]: E1004 11:15:47.412926 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:15:49 crc kubenswrapper[5025]: I1004 11:15:49.643908 5025 scope.go:117] "RemoveContainer" containerID="326416e575dd0cdbb619a18fa4bb546727bf66104c9e149132bf3c2020c463cf" Oct 04 11:16:02 crc kubenswrapper[5025]: I1004 11:16:02.417196 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:16:02 crc kubenswrapper[5025]: E1004 11:16:02.417830 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:16:17 crc kubenswrapper[5025]: I1004 11:16:17.412912 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:16:17 crc kubenswrapper[5025]: E1004 11:16:17.414004 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:16:28 crc kubenswrapper[5025]: I1004 11:16:28.411648 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:16:28 crc kubenswrapper[5025]: E1004 11:16:28.413474 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:16:39 crc kubenswrapper[5025]: I1004 11:16:39.412095 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:16:39 crc kubenswrapper[5025]: E1004 11:16:39.415008 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:16:52 crc kubenswrapper[5025]: I1004 11:16:52.424497 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:16:52 crc kubenswrapper[5025]: E1004 11:16:52.425995 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:17:03 crc kubenswrapper[5025]: I1004 11:17:03.410930 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:17:03 crc kubenswrapper[5025]: E1004 11:17:03.411642 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:17:03 crc kubenswrapper[5025]: I1004 11:17:03.713696 5025 generic.go:334] "Generic (PLEG): container finished" podID="d807eca9-05be-475e-b168-2ab47ab555c1" containerID="4ecac9300bb6a7d9a0eb131034b8f54be13474f7775151be661f12664adc8a06" exitCode=0 Oct 04 11:17:03 crc kubenswrapper[5025]: I1004 11:17:03.713787 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" event={"ID":"d807eca9-05be-475e-b168-2ab47ab555c1","Type":"ContainerDied","Data":"4ecac9300bb6a7d9a0eb131034b8f54be13474f7775151be661f12664adc8a06"} Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.146753 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.289728 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-ssh-key\") pod \"d807eca9-05be-475e-b168-2ab47ab555c1\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.289896 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-1\") pod \"d807eca9-05be-475e-b168-2ab47ab555c1\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.289953 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d807eca9-05be-475e-b168-2ab47ab555c1-nova-extra-config-0\") pod \"d807eca9-05be-475e-b168-2ab47ab555c1\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.290078 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-0\") pod \"d807eca9-05be-475e-b168-2ab47ab555c1\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.290237 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp677\" (UniqueName: \"kubernetes.io/projected/d807eca9-05be-475e-b168-2ab47ab555c1-kube-api-access-dp677\") pod \"d807eca9-05be-475e-b168-2ab47ab555c1\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.290277 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-inventory\") pod \"d807eca9-05be-475e-b168-2ab47ab555c1\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.290979 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-combined-ca-bundle\") pod \"d807eca9-05be-475e-b168-2ab47ab555c1\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.291214 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-0\") pod \"d807eca9-05be-475e-b168-2ab47ab555c1\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.291296 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-1\") pod \"d807eca9-05be-475e-b168-2ab47ab555c1\" (UID: \"d807eca9-05be-475e-b168-2ab47ab555c1\") " Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.299173 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d807eca9-05be-475e-b168-2ab47ab555c1-kube-api-access-dp677" (OuterVolumeSpecName: "kube-api-access-dp677") pod "d807eca9-05be-475e-b168-2ab47ab555c1" (UID: "d807eca9-05be-475e-b168-2ab47ab555c1"). InnerVolumeSpecName "kube-api-access-dp677". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.303856 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d807eca9-05be-475e-b168-2ab47ab555c1" (UID: "d807eca9-05be-475e-b168-2ab47ab555c1"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.336549 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d807eca9-05be-475e-b168-2ab47ab555c1-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "d807eca9-05be-475e-b168-2ab47ab555c1" (UID: "d807eca9-05be-475e-b168-2ab47ab555c1"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.337881 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "d807eca9-05be-475e-b168-2ab47ab555c1" (UID: "d807eca9-05be-475e-b168-2ab47ab555c1"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.341025 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "d807eca9-05be-475e-b168-2ab47ab555c1" (UID: "d807eca9-05be-475e-b168-2ab47ab555c1"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.342006 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-inventory" (OuterVolumeSpecName: "inventory") pod "d807eca9-05be-475e-b168-2ab47ab555c1" (UID: "d807eca9-05be-475e-b168-2ab47ab555c1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.348186 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d807eca9-05be-475e-b168-2ab47ab555c1" (UID: "d807eca9-05be-475e-b168-2ab47ab555c1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.349729 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "d807eca9-05be-475e-b168-2ab47ab555c1" (UID: "d807eca9-05be-475e-b168-2ab47ab555c1"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.354262 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "d807eca9-05be-475e-b168-2ab47ab555c1" (UID: "d807eca9-05be-475e-b168-2ab47ab555c1"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.395164 5025 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.395204 5025 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.395218 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.395230 5025 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/d807eca9-05be-475e-b168-2ab47ab555c1-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.395242 5025 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.395255 5025 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.395267 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp677\" (UniqueName: \"kubernetes.io/projected/d807eca9-05be-475e-b168-2ab47ab555c1-kube-api-access-dp677\") on node \"crc\" DevicePath \"\"" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.395278 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.395290 5025 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d807eca9-05be-475e-b168-2ab47ab555c1-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.732306 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" event={"ID":"d807eca9-05be-475e-b168-2ab47ab555c1","Type":"ContainerDied","Data":"9a296e0ddfc6ceda2173cb11ad9f88b42eeacf839235c1398e13d98032b4d231"} Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.732350 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a296e0ddfc6ceda2173cb11ad9f88b42eeacf839235c1398e13d98032b4d231" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.732360 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-tqj97" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.867126 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd"] Oct 04 11:17:05 crc kubenswrapper[5025]: E1004 11:17:05.867933 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d807eca9-05be-475e-b168-2ab47ab555c1" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.867972 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="d807eca9-05be-475e-b168-2ab47ab555c1" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 11:17:05 crc kubenswrapper[5025]: E1004 11:17:05.868170 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9332a111-c0da-4fc5-b9ee-462ff503750e" containerName="collect-profiles" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.868201 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9332a111-c0da-4fc5-b9ee-462ff503750e" containerName="collect-profiles" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.868643 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="d807eca9-05be-475e-b168-2ab47ab555c1" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.868720 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="9332a111-c0da-4fc5-b9ee-462ff503750e" containerName="collect-profiles" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.870080 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.872567 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.872859 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.872740 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-dntlh" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.873062 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.873372 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.875458 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd"] Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.904584 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvh5n\" (UniqueName: \"kubernetes.io/projected/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-kube-api-access-fvh5n\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.904652 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.904736 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.904777 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.904883 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.904920 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:05 crc kubenswrapper[5025]: I1004 11:17:05.904954 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.006506 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.006556 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.006642 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.006698 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.006724 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.006812 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvh5n\" (UniqueName: \"kubernetes.io/projected/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-kube-api-access-fvh5n\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.006842 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.012384 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.012401 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.012840 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.021156 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.021691 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.022288 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.025123 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvh5n\" (UniqueName: \"kubernetes.io/projected/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-kube-api-access-fvh5n\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.198612 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:17:06 crc kubenswrapper[5025]: I1004 11:17:06.737732 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd"] Oct 04 11:17:07 crc kubenswrapper[5025]: I1004 11:17:07.753272 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" event={"ID":"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f","Type":"ContainerStarted","Data":"64461ca4f53b4181b91ee25b26a4943c214e0747e26bc7b9ce93f9e18f596364"} Oct 04 11:17:07 crc kubenswrapper[5025]: I1004 11:17:07.753636 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" event={"ID":"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f","Type":"ContainerStarted","Data":"ea7d31cd5d980ca4e0768d17be841efda91b8170614dfab7890e9bf0b51cb1f0"} Oct 04 11:17:18 crc kubenswrapper[5025]: I1004 11:17:18.411551 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:17:18 crc kubenswrapper[5025]: E1004 11:17:18.412321 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:17:32 crc kubenswrapper[5025]: I1004 11:17:32.422296 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:17:32 crc kubenswrapper[5025]: E1004 11:17:32.425359 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:17:47 crc kubenswrapper[5025]: I1004 11:17:47.412746 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:17:47 crc kubenswrapper[5025]: E1004 11:17:47.414082 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:18:00 crc kubenswrapper[5025]: I1004 11:18:00.411387 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:18:00 crc kubenswrapper[5025]: E1004 11:18:00.412237 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:18:15 crc kubenswrapper[5025]: I1004 11:18:15.411202 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:18:15 crc kubenswrapper[5025]: E1004 11:18:15.412146 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:18:27 crc kubenswrapper[5025]: I1004 11:18:27.412407 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:18:27 crc kubenswrapper[5025]: E1004 11:18:27.414578 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.432591 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" podStartSLOduration=84.943950626 podStartE2EDuration="1m25.432546842s" podCreationTimestamp="2025-10-04 11:17:05 +0000 UTC" firstStartedPulling="2025-10-04 11:17:06.742915984 +0000 UTC m=+2555.167882874" lastFinishedPulling="2025-10-04 11:17:07.23151219 +0000 UTC m=+2555.656479090" observedRunningTime="2025-10-04 11:17:07.779234399 +0000 UTC m=+2556.204201359" watchObservedRunningTime="2025-10-04 11:18:30.432546842 +0000 UTC m=+2638.857513772" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.443776 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-859l2"] Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.448195 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.456226 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-859l2"] Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.531957 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbs64\" (UniqueName: \"kubernetes.io/projected/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-kube-api-access-dbs64\") pod \"redhat-marketplace-859l2\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.532241 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-catalog-content\") pod \"redhat-marketplace-859l2\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.532305 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-utilities\") pod \"redhat-marketplace-859l2\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.633947 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-catalog-content\") pod \"redhat-marketplace-859l2\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.634054 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-utilities\") pod \"redhat-marketplace-859l2\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.634100 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbs64\" (UniqueName: \"kubernetes.io/projected/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-kube-api-access-dbs64\") pod \"redhat-marketplace-859l2\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.634500 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-catalog-content\") pod \"redhat-marketplace-859l2\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.634570 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-utilities\") pod \"redhat-marketplace-859l2\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.661400 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbs64\" (UniqueName: \"kubernetes.io/projected/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-kube-api-access-dbs64\") pod \"redhat-marketplace-859l2\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:30 crc kubenswrapper[5025]: I1004 11:18:30.780847 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:31 crc kubenswrapper[5025]: I1004 11:18:31.252276 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-859l2"] Oct 04 11:18:31 crc kubenswrapper[5025]: I1004 11:18:31.627316 5025 generic.go:334] "Generic (PLEG): container finished" podID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerID="4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069" exitCode=0 Oct 04 11:18:31 crc kubenswrapper[5025]: I1004 11:18:31.627372 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-859l2" event={"ID":"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef","Type":"ContainerDied","Data":"4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069"} Oct 04 11:18:31 crc kubenswrapper[5025]: I1004 11:18:31.627425 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-859l2" event={"ID":"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef","Type":"ContainerStarted","Data":"9073c0fc7ac3cd40b4712ec5d6978c5041676bfec4c8f5464ddc418429754b42"} Oct 04 11:18:32 crc kubenswrapper[5025]: I1004 11:18:32.638477 5025 generic.go:334] "Generic (PLEG): container finished" podID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerID="9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620" exitCode=0 Oct 04 11:18:32 crc kubenswrapper[5025]: I1004 11:18:32.638584 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-859l2" event={"ID":"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef","Type":"ContainerDied","Data":"9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620"} Oct 04 11:18:33 crc kubenswrapper[5025]: I1004 11:18:33.654890 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-859l2" event={"ID":"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef","Type":"ContainerStarted","Data":"d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9"} Oct 04 11:18:33 crc kubenswrapper[5025]: I1004 11:18:33.684669 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-859l2" podStartSLOduration=2.264772667 podStartE2EDuration="3.684644768s" podCreationTimestamp="2025-10-04 11:18:30 +0000 UTC" firstStartedPulling="2025-10-04 11:18:31.630601692 +0000 UTC m=+2640.055568582" lastFinishedPulling="2025-10-04 11:18:33.050473773 +0000 UTC m=+2641.475440683" observedRunningTime="2025-10-04 11:18:33.671977089 +0000 UTC m=+2642.096943979" watchObservedRunningTime="2025-10-04 11:18:33.684644768 +0000 UTC m=+2642.109611668" Oct 04 11:18:40 crc kubenswrapper[5025]: I1004 11:18:40.781685 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:40 crc kubenswrapper[5025]: I1004 11:18:40.782396 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:40 crc kubenswrapper[5025]: I1004 11:18:40.870546 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:41 crc kubenswrapper[5025]: I1004 11:18:41.783697 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:41 crc kubenswrapper[5025]: I1004 11:18:41.837340 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-859l2"] Oct 04 11:18:42 crc kubenswrapper[5025]: I1004 11:18:42.420571 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:18:42 crc kubenswrapper[5025]: E1004 11:18:42.420833 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:18:43 crc kubenswrapper[5025]: I1004 11:18:43.753961 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-859l2" podUID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerName="registry-server" containerID="cri-o://d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9" gracePeriod=2 Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.245151 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.296542 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-utilities\") pod \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.297187 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-catalog-content\") pod \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.297393 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbs64\" (UniqueName: \"kubernetes.io/projected/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-kube-api-access-dbs64\") pod \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\" (UID: \"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef\") " Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.297616 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-utilities" (OuterVolumeSpecName: "utilities") pod "3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" (UID: "3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.297868 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.302679 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-kube-api-access-dbs64" (OuterVolumeSpecName: "kube-api-access-dbs64") pod "3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" (UID: "3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef"). InnerVolumeSpecName "kube-api-access-dbs64". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.308991 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" (UID: "3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.399436 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.399472 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbs64\" (UniqueName: \"kubernetes.io/projected/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef-kube-api-access-dbs64\") on node \"crc\" DevicePath \"\"" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.769432 5025 generic.go:334] "Generic (PLEG): container finished" podID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerID="d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9" exitCode=0 Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.769474 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-859l2" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.769576 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-859l2" event={"ID":"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef","Type":"ContainerDied","Data":"d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9"} Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.769629 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-859l2" event={"ID":"3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef","Type":"ContainerDied","Data":"9073c0fc7ac3cd40b4712ec5d6978c5041676bfec4c8f5464ddc418429754b42"} Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.769660 5025 scope.go:117] "RemoveContainer" containerID="d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.798501 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-859l2"] Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.805490 5025 scope.go:117] "RemoveContainer" containerID="9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.808184 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-859l2"] Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.839382 5025 scope.go:117] "RemoveContainer" containerID="4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.929891 5025 scope.go:117] "RemoveContainer" containerID="d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9" Oct 04 11:18:44 crc kubenswrapper[5025]: E1004 11:18:44.930760 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9\": container with ID starting with d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9 not found: ID does not exist" containerID="d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.930821 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9"} err="failed to get container status \"d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9\": rpc error: code = NotFound desc = could not find container \"d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9\": container with ID starting with d68d49a60b655a67f0f549e7b3d2c292e90edd4c95a4550599431b9a8215cde9 not found: ID does not exist" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.930852 5025 scope.go:117] "RemoveContainer" containerID="9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620" Oct 04 11:18:44 crc kubenswrapper[5025]: E1004 11:18:44.931402 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620\": container with ID starting with 9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620 not found: ID does not exist" containerID="9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.931507 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620"} err="failed to get container status \"9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620\": rpc error: code = NotFound desc = could not find container \"9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620\": container with ID starting with 9adc9c8bc5187696a50e62538987cd44ff0ccb900a6c4163f492a63c4be01620 not found: ID does not exist" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.931614 5025 scope.go:117] "RemoveContainer" containerID="4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069" Oct 04 11:18:44 crc kubenswrapper[5025]: E1004 11:18:44.932068 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069\": container with ID starting with 4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069 not found: ID does not exist" containerID="4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069" Oct 04 11:18:44 crc kubenswrapper[5025]: I1004 11:18:44.932122 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069"} err="failed to get container status \"4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069\": rpc error: code = NotFound desc = could not find container \"4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069\": container with ID starting with 4718ca147d9586262afb2404377e34b2dca9145d69293ce5eb5b175e4d73d069 not found: ID does not exist" Oct 04 11:18:46 crc kubenswrapper[5025]: I1004 11:18:46.424677 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" path="/var/lib/kubelet/pods/3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef/volumes" Oct 04 11:18:55 crc kubenswrapper[5025]: I1004 11:18:55.411885 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:18:55 crc kubenswrapper[5025]: E1004 11:18:55.412982 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:19:07 crc kubenswrapper[5025]: I1004 11:19:07.412213 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:19:07 crc kubenswrapper[5025]: E1004 11:19:07.413641 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:19:21 crc kubenswrapper[5025]: I1004 11:19:21.412515 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:19:22 crc kubenswrapper[5025]: I1004 11:19:22.233572 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"6ddf274783bd9ad4cb17c313083966159dc88be7cc9e9b27f856feb0a8e136a6"} Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.671353 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pgcjs"] Oct 04 11:19:24 crc kubenswrapper[5025]: E1004 11:19:24.673367 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerName="extract-utilities" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.673472 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerName="extract-utilities" Oct 04 11:19:24 crc kubenswrapper[5025]: E1004 11:19:24.673567 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerName="registry-server" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.673641 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerName="registry-server" Oct 04 11:19:24 crc kubenswrapper[5025]: E1004 11:19:24.673753 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerName="extract-content" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.673831 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerName="extract-content" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.674143 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d9c8bb7-e6e9-4eb6-b797-267dfc58ceef" containerName="registry-server" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.675924 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.683729 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pgcjs"] Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.829326 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-catalog-content\") pod \"redhat-operators-pgcjs\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.829412 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8vbb\" (UniqueName: \"kubernetes.io/projected/9d129f00-0219-41e0-b375-a7bbd9872f37-kube-api-access-n8vbb\") pod \"redhat-operators-pgcjs\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.829457 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-utilities\") pod \"redhat-operators-pgcjs\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.930940 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8vbb\" (UniqueName: \"kubernetes.io/projected/9d129f00-0219-41e0-b375-a7bbd9872f37-kube-api-access-n8vbb\") pod \"redhat-operators-pgcjs\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.931031 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-utilities\") pod \"redhat-operators-pgcjs\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.931151 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-catalog-content\") pod \"redhat-operators-pgcjs\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.931620 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-utilities\") pod \"redhat-operators-pgcjs\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.931684 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-catalog-content\") pod \"redhat-operators-pgcjs\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:24 crc kubenswrapper[5025]: I1004 11:19:24.956092 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8vbb\" (UniqueName: \"kubernetes.io/projected/9d129f00-0219-41e0-b375-a7bbd9872f37-kube-api-access-n8vbb\") pod \"redhat-operators-pgcjs\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:25 crc kubenswrapper[5025]: I1004 11:19:25.003330 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:25 crc kubenswrapper[5025]: I1004 11:19:25.489069 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pgcjs"] Oct 04 11:19:26 crc kubenswrapper[5025]: I1004 11:19:26.281833 5025 generic.go:334] "Generic (PLEG): container finished" podID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerID="36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900" exitCode=0 Oct 04 11:19:26 crc kubenswrapper[5025]: I1004 11:19:26.282025 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgcjs" event={"ID":"9d129f00-0219-41e0-b375-a7bbd9872f37","Type":"ContainerDied","Data":"36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900"} Oct 04 11:19:26 crc kubenswrapper[5025]: I1004 11:19:26.282230 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgcjs" event={"ID":"9d129f00-0219-41e0-b375-a7bbd9872f37","Type":"ContainerStarted","Data":"ac7f1847dbac72b0eaf285d0f0b4f29f91f363227229cb9cc04670179956abec"} Oct 04 11:19:26 crc kubenswrapper[5025]: I1004 11:19:26.285099 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:19:28 crc kubenswrapper[5025]: I1004 11:19:28.319223 5025 generic.go:334] "Generic (PLEG): container finished" podID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerID="e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3" exitCode=0 Oct 04 11:19:28 crc kubenswrapper[5025]: I1004 11:19:28.319311 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgcjs" event={"ID":"9d129f00-0219-41e0-b375-a7bbd9872f37","Type":"ContainerDied","Data":"e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3"} Oct 04 11:19:29 crc kubenswrapper[5025]: I1004 11:19:29.337629 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgcjs" event={"ID":"9d129f00-0219-41e0-b375-a7bbd9872f37","Type":"ContainerStarted","Data":"a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2"} Oct 04 11:19:35 crc kubenswrapper[5025]: I1004 11:19:35.003927 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:35 crc kubenswrapper[5025]: I1004 11:19:35.004646 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:35 crc kubenswrapper[5025]: I1004 11:19:35.090068 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:35 crc kubenswrapper[5025]: I1004 11:19:35.120666 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pgcjs" podStartSLOduration=8.668150821 podStartE2EDuration="11.12064597s" podCreationTimestamp="2025-10-04 11:19:24 +0000 UTC" firstStartedPulling="2025-10-04 11:19:26.284661278 +0000 UTC m=+2694.709628198" lastFinishedPulling="2025-10-04 11:19:28.737156427 +0000 UTC m=+2697.162123347" observedRunningTime="2025-10-04 11:19:29.369643291 +0000 UTC m=+2697.794610201" watchObservedRunningTime="2025-10-04 11:19:35.12064597 +0000 UTC m=+2703.545612860" Oct 04 11:19:35 crc kubenswrapper[5025]: I1004 11:19:35.484850 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:35 crc kubenswrapper[5025]: I1004 11:19:35.544439 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pgcjs"] Oct 04 11:19:37 crc kubenswrapper[5025]: I1004 11:19:37.434547 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pgcjs" podUID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerName="registry-server" containerID="cri-o://a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2" gracePeriod=2 Oct 04 11:19:37 crc kubenswrapper[5025]: I1004 11:19:37.943159 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.107337 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8vbb\" (UniqueName: \"kubernetes.io/projected/9d129f00-0219-41e0-b375-a7bbd9872f37-kube-api-access-n8vbb\") pod \"9d129f00-0219-41e0-b375-a7bbd9872f37\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.107476 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-utilities\") pod \"9d129f00-0219-41e0-b375-a7bbd9872f37\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.107556 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-catalog-content\") pod \"9d129f00-0219-41e0-b375-a7bbd9872f37\" (UID: \"9d129f00-0219-41e0-b375-a7bbd9872f37\") " Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.108815 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-utilities" (OuterVolumeSpecName: "utilities") pod "9d129f00-0219-41e0-b375-a7bbd9872f37" (UID: "9d129f00-0219-41e0-b375-a7bbd9872f37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.116152 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d129f00-0219-41e0-b375-a7bbd9872f37-kube-api-access-n8vbb" (OuterVolumeSpecName: "kube-api-access-n8vbb") pod "9d129f00-0219-41e0-b375-a7bbd9872f37" (UID: "9d129f00-0219-41e0-b375-a7bbd9872f37"). InnerVolumeSpecName "kube-api-access-n8vbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.210539 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8vbb\" (UniqueName: \"kubernetes.io/projected/9d129f00-0219-41e0-b375-a7bbd9872f37-kube-api-access-n8vbb\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.210863 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.447786 5025 generic.go:334] "Generic (PLEG): container finished" podID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerID="a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2" exitCode=0 Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.447876 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgcjs" event={"ID":"9d129f00-0219-41e0-b375-a7bbd9872f37","Type":"ContainerDied","Data":"a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2"} Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.447919 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pgcjs" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.447940 5025 scope.go:117] "RemoveContainer" containerID="a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.447927 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pgcjs" event={"ID":"9d129f00-0219-41e0-b375-a7bbd9872f37","Type":"ContainerDied","Data":"ac7f1847dbac72b0eaf285d0f0b4f29f91f363227229cb9cc04670179956abec"} Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.480354 5025 scope.go:117] "RemoveContainer" containerID="e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.525382 5025 scope.go:117] "RemoveContainer" containerID="36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.571313 5025 scope.go:117] "RemoveContainer" containerID="a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2" Oct 04 11:19:38 crc kubenswrapper[5025]: E1004 11:19:38.571944 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2\": container with ID starting with a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2 not found: ID does not exist" containerID="a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.571990 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2"} err="failed to get container status \"a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2\": rpc error: code = NotFound desc = could not find container \"a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2\": container with ID starting with a57d377677fa1f913b497a697625b6e253a8c7834aaafdf46cefeae6790c2ff2 not found: ID does not exist" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.572035 5025 scope.go:117] "RemoveContainer" containerID="e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3" Oct 04 11:19:38 crc kubenswrapper[5025]: E1004 11:19:38.572455 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3\": container with ID starting with e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3 not found: ID does not exist" containerID="e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.572492 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3"} err="failed to get container status \"e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3\": rpc error: code = NotFound desc = could not find container \"e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3\": container with ID starting with e17dc28304e3f4c388386cbd540fb3c28ed7b915b6b2657db346461cb159bfd3 not found: ID does not exist" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.572518 5025 scope.go:117] "RemoveContainer" containerID="36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900" Oct 04 11:19:38 crc kubenswrapper[5025]: E1004 11:19:38.572822 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900\": container with ID starting with 36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900 not found: ID does not exist" containerID="36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900" Oct 04 11:19:38 crc kubenswrapper[5025]: I1004 11:19:38.572863 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900"} err="failed to get container status \"36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900\": rpc error: code = NotFound desc = could not find container \"36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900\": container with ID starting with 36bd25b4a29ec32714a99f821ec9bc7015fe717b3afc8439262eb7dfe376e900 not found: ID does not exist" Oct 04 11:19:39 crc kubenswrapper[5025]: I1004 11:19:39.451926 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d129f00-0219-41e0-b375-a7bbd9872f37" (UID: "9d129f00-0219-41e0-b375-a7bbd9872f37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:19:39 crc kubenswrapper[5025]: I1004 11:19:39.537386 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d129f00-0219-41e0-b375-a7bbd9872f37-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:39 crc kubenswrapper[5025]: I1004 11:19:39.696317 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pgcjs"] Oct 04 11:19:39 crc kubenswrapper[5025]: I1004 11:19:39.705042 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pgcjs"] Oct 04 11:19:40 crc kubenswrapper[5025]: I1004 11:19:40.424237 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d129f00-0219-41e0-b375-a7bbd9872f37" path="/var/lib/kubelet/pods/9d129f00-0219-41e0-b375-a7bbd9872f37/volumes" Oct 04 11:19:41 crc kubenswrapper[5025]: I1004 11:19:41.488455 5025 generic.go:334] "Generic (PLEG): container finished" podID="4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" containerID="64461ca4f53b4181b91ee25b26a4943c214e0747e26bc7b9ce93f9e18f596364" exitCode=0 Oct 04 11:19:41 crc kubenswrapper[5025]: I1004 11:19:41.488513 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" event={"ID":"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f","Type":"ContainerDied","Data":"64461ca4f53b4181b91ee25b26a4943c214e0747e26bc7b9ce93f9e18f596364"} Oct 04 11:19:42 crc kubenswrapper[5025]: I1004 11:19:42.944307 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.113575 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-inventory\") pod \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.113625 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-1\") pod \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.113697 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-2\") pod \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.113715 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-telemetry-combined-ca-bundle\") pod \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.113771 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-0\") pod \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.113809 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvh5n\" (UniqueName: \"kubernetes.io/projected/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-kube-api-access-fvh5n\") pod \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.113830 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ssh-key\") pod \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\" (UID: \"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f\") " Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.118784 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" (UID: "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.119399 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-kube-api-access-fvh5n" (OuterVolumeSpecName: "kube-api-access-fvh5n") pod "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" (UID: "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f"). InnerVolumeSpecName "kube-api-access-fvh5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.140283 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" (UID: "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.141742 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-inventory" (OuterVolumeSpecName: "inventory") pod "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" (UID: "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.147673 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" (UID: "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.157129 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" (UID: "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.160363 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" (UID: "4d9059c6-5fa0-4cc7-a7b5-490f85405f9f"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.216064 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvh5n\" (UniqueName: \"kubernetes.io/projected/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-kube-api-access-fvh5n\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.216095 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.216104 5025 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-inventory\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.216116 5025 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.216125 5025 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.216134 5025 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.216143 5025 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d9059c6-5fa0-4cc7-a7b5-490f85405f9f-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.521818 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" event={"ID":"4d9059c6-5fa0-4cc7-a7b5-490f85405f9f","Type":"ContainerDied","Data":"ea7d31cd5d980ca4e0768d17be841efda91b8170614dfab7890e9bf0b51cb1f0"} Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.521880 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea7d31cd5d980ca4e0768d17be841efda91b8170614dfab7890e9bf0b51cb1f0" Oct 04 11:19:43 crc kubenswrapper[5025]: I1004 11:19:43.521921 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.149080 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 11:20:36 crc kubenswrapper[5025]: E1004 11:20:36.150592 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerName="extract-utilities" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.150618 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerName="extract-utilities" Oct 04 11:20:36 crc kubenswrapper[5025]: E1004 11:20:36.150645 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerName="registry-server" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.150655 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerName="registry-server" Oct 04 11:20:36 crc kubenswrapper[5025]: E1004 11:20:36.150679 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.150690 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 11:20:36 crc kubenswrapper[5025]: E1004 11:20:36.150705 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerName="extract-content" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.150715 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerName="extract-content" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.151084 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d129f00-0219-41e0-b375-a7bbd9872f37" containerName="registry-server" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.151110 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9059c6-5fa0-4cc7-a7b5-490f85405f9f" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.152717 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.154886 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.154971 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.155079 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.156160 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-bdqc5" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.173561 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.246927 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzzgs\" (UniqueName: \"kubernetes.io/projected/5c783e4c-998b-44ab-956a-de26d5568f90-kube-api-access-hzzgs\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.247027 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.247361 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.247549 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-config-data\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.247616 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.247694 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.247738 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.247788 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.247821 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.350795 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.350877 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-config-data\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.350907 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.350940 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.350968 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.351029 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.351064 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.351108 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzzgs\" (UniqueName: \"kubernetes.io/projected/5c783e4c-998b-44ab-956a-de26d5568f90-kube-api-access-hzzgs\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.351199 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.351912 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.351914 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.352408 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.353493 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.353677 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-config-data\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.361337 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.362583 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.367370 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.376465 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzzgs\" (UniqueName: \"kubernetes.io/projected/5c783e4c-998b-44ab-956a-de26d5568f90-kube-api-access-hzzgs\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.396046 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.483546 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 11:20:36 crc kubenswrapper[5025]: I1004 11:20:36.947412 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 04 11:20:37 crc kubenswrapper[5025]: I1004 11:20:37.136317 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5c783e4c-998b-44ab-956a-de26d5568f90","Type":"ContainerStarted","Data":"1477d226cc55bb59c5d7a84a91872a12f4a9e479593e9257c60379cd7ba4c507"} Oct 04 11:21:04 crc kubenswrapper[5025]: E1004 11:21:04.412913 5025 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 04 11:21:04 crc kubenswrapper[5025]: E1004 11:21:04.413941 5025 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hzzgs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(5c783e4c-998b-44ab-956a-de26d5568f90): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 04 11:21:04 crc kubenswrapper[5025]: E1004 11:21:04.420568 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="5c783e4c-998b-44ab-956a-de26d5568f90" Oct 04 11:21:05 crc kubenswrapper[5025]: E1004 11:21:05.402682 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="5c783e4c-998b-44ab-956a-de26d5568f90" Oct 04 11:21:07 crc kubenswrapper[5025]: I1004 11:21:07.970171 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m6hmv"] Oct 04 11:21:07 crc kubenswrapper[5025]: I1004 11:21:07.972932 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:07 crc kubenswrapper[5025]: I1004 11:21:07.980521 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m6hmv"] Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.019801 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-catalog-content\") pod \"certified-operators-m6hmv\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.019969 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-utilities\") pod \"certified-operators-m6hmv\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.020077 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6plq\" (UniqueName: \"kubernetes.io/projected/d0fc8628-5c4b-418a-abed-32c047c1eb72-kube-api-access-w6plq\") pod \"certified-operators-m6hmv\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.122556 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-utilities\") pod \"certified-operators-m6hmv\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.122673 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6plq\" (UniqueName: \"kubernetes.io/projected/d0fc8628-5c4b-418a-abed-32c047c1eb72-kube-api-access-w6plq\") pod \"certified-operators-m6hmv\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.122962 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-catalog-content\") pod \"certified-operators-m6hmv\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.123538 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-catalog-content\") pod \"certified-operators-m6hmv\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.123750 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-utilities\") pod \"certified-operators-m6hmv\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.143059 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6plq\" (UniqueName: \"kubernetes.io/projected/d0fc8628-5c4b-418a-abed-32c047c1eb72-kube-api-access-w6plq\") pod \"certified-operators-m6hmv\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.296040 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:08 crc kubenswrapper[5025]: I1004 11:21:08.796399 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m6hmv"] Oct 04 11:21:09 crc kubenswrapper[5025]: I1004 11:21:09.447339 5025 generic.go:334] "Generic (PLEG): container finished" podID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerID="406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e" exitCode=0 Oct 04 11:21:09 crc kubenswrapper[5025]: I1004 11:21:09.447461 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6hmv" event={"ID":"d0fc8628-5c4b-418a-abed-32c047c1eb72","Type":"ContainerDied","Data":"406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e"} Oct 04 11:21:09 crc kubenswrapper[5025]: I1004 11:21:09.447711 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6hmv" event={"ID":"d0fc8628-5c4b-418a-abed-32c047c1eb72","Type":"ContainerStarted","Data":"6e75a694ba66cdb1f40230bf2e11d460d80090019aa3b856d4dd0ae612b35872"} Oct 04 11:21:10 crc kubenswrapper[5025]: I1004 11:21:10.462293 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6hmv" event={"ID":"d0fc8628-5c4b-418a-abed-32c047c1eb72","Type":"ContainerStarted","Data":"265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d"} Oct 04 11:21:11 crc kubenswrapper[5025]: I1004 11:21:11.482642 5025 generic.go:334] "Generic (PLEG): container finished" podID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerID="265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d" exitCode=0 Oct 04 11:21:11 crc kubenswrapper[5025]: I1004 11:21:11.482723 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6hmv" event={"ID":"d0fc8628-5c4b-418a-abed-32c047c1eb72","Type":"ContainerDied","Data":"265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d"} Oct 04 11:21:12 crc kubenswrapper[5025]: I1004 11:21:12.492200 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6hmv" event={"ID":"d0fc8628-5c4b-418a-abed-32c047c1eb72","Type":"ContainerStarted","Data":"ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f"} Oct 04 11:21:12 crc kubenswrapper[5025]: I1004 11:21:12.516178 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m6hmv" podStartSLOduration=3.078254474 podStartE2EDuration="5.51615523s" podCreationTimestamp="2025-10-04 11:21:07 +0000 UTC" firstStartedPulling="2025-10-04 11:21:09.449256221 +0000 UTC m=+2797.874223131" lastFinishedPulling="2025-10-04 11:21:11.887157007 +0000 UTC m=+2800.312123887" observedRunningTime="2025-10-04 11:21:12.510662911 +0000 UTC m=+2800.935629831" watchObservedRunningTime="2025-10-04 11:21:12.51615523 +0000 UTC m=+2800.941122150" Oct 04 11:21:18 crc kubenswrapper[5025]: I1004 11:21:18.296476 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:18 crc kubenswrapper[5025]: I1004 11:21:18.296776 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:18 crc kubenswrapper[5025]: I1004 11:21:18.365046 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:18 crc kubenswrapper[5025]: I1004 11:21:18.632088 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:18 crc kubenswrapper[5025]: I1004 11:21:18.687695 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m6hmv"] Oct 04 11:21:19 crc kubenswrapper[5025]: I1004 11:21:19.571515 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5c783e4c-998b-44ab-956a-de26d5568f90","Type":"ContainerStarted","Data":"0abe21bb6337ddcfb19fd7f1b1d8a78ef17f4cf75fcd6542f3e7ecb835320382"} Oct 04 11:21:19 crc kubenswrapper[5025]: I1004 11:21:19.607932 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.728290043 podStartE2EDuration="44.607913217s" podCreationTimestamp="2025-10-04 11:20:35 +0000 UTC" firstStartedPulling="2025-10-04 11:20:36.957288104 +0000 UTC m=+2765.382254984" lastFinishedPulling="2025-10-04 11:21:17.836911278 +0000 UTC m=+2806.261878158" observedRunningTime="2025-10-04 11:21:19.5917677 +0000 UTC m=+2808.016734590" watchObservedRunningTime="2025-10-04 11:21:19.607913217 +0000 UTC m=+2808.032880097" Oct 04 11:21:20 crc kubenswrapper[5025]: I1004 11:21:20.581952 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m6hmv" podUID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerName="registry-server" containerID="cri-o://ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f" gracePeriod=2 Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.062637 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.111672 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6plq\" (UniqueName: \"kubernetes.io/projected/d0fc8628-5c4b-418a-abed-32c047c1eb72-kube-api-access-w6plq\") pod \"d0fc8628-5c4b-418a-abed-32c047c1eb72\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.111997 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-catalog-content\") pod \"d0fc8628-5c4b-418a-abed-32c047c1eb72\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.112189 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-utilities\") pod \"d0fc8628-5c4b-418a-abed-32c047c1eb72\" (UID: \"d0fc8628-5c4b-418a-abed-32c047c1eb72\") " Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.114426 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-utilities" (OuterVolumeSpecName: "utilities") pod "d0fc8628-5c4b-418a-abed-32c047c1eb72" (UID: "d0fc8628-5c4b-418a-abed-32c047c1eb72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.119597 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0fc8628-5c4b-418a-abed-32c047c1eb72-kube-api-access-w6plq" (OuterVolumeSpecName: "kube-api-access-w6plq") pod "d0fc8628-5c4b-418a-abed-32c047c1eb72" (UID: "d0fc8628-5c4b-418a-abed-32c047c1eb72"). InnerVolumeSpecName "kube-api-access-w6plq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.214686 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6plq\" (UniqueName: \"kubernetes.io/projected/d0fc8628-5c4b-418a-abed-32c047c1eb72-kube-api-access-w6plq\") on node \"crc\" DevicePath \"\"" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.214713 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.593678 5025 generic.go:334] "Generic (PLEG): container finished" podID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerID="ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f" exitCode=0 Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.593714 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6hmv" event={"ID":"d0fc8628-5c4b-418a-abed-32c047c1eb72","Type":"ContainerDied","Data":"ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f"} Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.593738 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m6hmv" event={"ID":"d0fc8628-5c4b-418a-abed-32c047c1eb72","Type":"ContainerDied","Data":"6e75a694ba66cdb1f40230bf2e11d460d80090019aa3b856d4dd0ae612b35872"} Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.593757 5025 scope.go:117] "RemoveContainer" containerID="ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.595359 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m6hmv" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.615836 5025 scope.go:117] "RemoveContainer" containerID="265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.644529 5025 scope.go:117] "RemoveContainer" containerID="406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.706186 5025 scope.go:117] "RemoveContainer" containerID="ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f" Oct 04 11:21:21 crc kubenswrapper[5025]: E1004 11:21:21.706720 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f\": container with ID starting with ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f not found: ID does not exist" containerID="ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.706760 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f"} err="failed to get container status \"ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f\": rpc error: code = NotFound desc = could not find container \"ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f\": container with ID starting with ce43b349ffebdab828fadaeb4afde33e830d42cc90c710358d9a4d71b696fd9f not found: ID does not exist" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.706787 5025 scope.go:117] "RemoveContainer" containerID="265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d" Oct 04 11:21:21 crc kubenswrapper[5025]: E1004 11:21:21.707307 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d\": container with ID starting with 265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d not found: ID does not exist" containerID="265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.707358 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d"} err="failed to get container status \"265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d\": rpc error: code = NotFound desc = could not find container \"265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d\": container with ID starting with 265fc63de2a6cc274dc8d68cede48626286ba82f3616fc702a29b94607242c6d not found: ID does not exist" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.707385 5025 scope.go:117] "RemoveContainer" containerID="406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e" Oct 04 11:21:21 crc kubenswrapper[5025]: E1004 11:21:21.707992 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e\": container with ID starting with 406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e not found: ID does not exist" containerID="406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.708066 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e"} err="failed to get container status \"406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e\": rpc error: code = NotFound desc = could not find container \"406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e\": container with ID starting with 406495524f384ef53d3e9e1dfad8fc8383f99496a0c823395574c7b4086d868e not found: ID does not exist" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.847221 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0fc8628-5c4b-418a-abed-32c047c1eb72" (UID: "d0fc8628-5c4b-418a-abed-32c047c1eb72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.927597 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0fc8628-5c4b-418a-abed-32c047c1eb72-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.939606 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m6hmv"] Oct 04 11:21:21 crc kubenswrapper[5025]: I1004 11:21:21.952975 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m6hmv"] Oct 04 11:21:22 crc kubenswrapper[5025]: I1004 11:21:22.438158 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0fc8628-5c4b-418a-abed-32c047c1eb72" path="/var/lib/kubelet/pods/d0fc8628-5c4b-418a-abed-32c047c1eb72/volumes" Oct 04 11:21:44 crc kubenswrapper[5025]: I1004 11:21:44.714099 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:21:44 crc kubenswrapper[5025]: I1004 11:21:44.714785 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:22:14 crc kubenswrapper[5025]: I1004 11:22:14.713420 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:22:14 crc kubenswrapper[5025]: I1004 11:22:14.714163 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:22:33 crc kubenswrapper[5025]: I1004 11:22:33.995821 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jkflp"] Oct 04 11:22:33 crc kubenswrapper[5025]: E1004 11:22:33.997650 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerName="extract-content" Oct 04 11:22:33 crc kubenswrapper[5025]: I1004 11:22:33.997676 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerName="extract-content" Oct 04 11:22:33 crc kubenswrapper[5025]: E1004 11:22:33.997729 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerName="extract-utilities" Oct 04 11:22:33 crc kubenswrapper[5025]: I1004 11:22:33.997749 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerName="extract-utilities" Oct 04 11:22:34 crc kubenswrapper[5025]: E1004 11:22:34.001254 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerName="registry-server" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.001349 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerName="registry-server" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.001960 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0fc8628-5c4b-418a-abed-32c047c1eb72" containerName="registry-server" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.004171 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.017367 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jkflp"] Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.081203 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-utilities\") pod \"community-operators-jkflp\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.081288 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc7z8\" (UniqueName: \"kubernetes.io/projected/3feb3cdb-d852-4499-8a6c-3d1f420612d7-kube-api-access-jc7z8\") pod \"community-operators-jkflp\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.081318 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-catalog-content\") pod \"community-operators-jkflp\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.184757 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-utilities\") pod \"community-operators-jkflp\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.184844 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc7z8\" (UniqueName: \"kubernetes.io/projected/3feb3cdb-d852-4499-8a6c-3d1f420612d7-kube-api-access-jc7z8\") pod \"community-operators-jkflp\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.184868 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-catalog-content\") pod \"community-operators-jkflp\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.185498 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-catalog-content\") pod \"community-operators-jkflp\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.185869 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-utilities\") pod \"community-operators-jkflp\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.213041 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc7z8\" (UniqueName: \"kubernetes.io/projected/3feb3cdb-d852-4499-8a6c-3d1f420612d7-kube-api-access-jc7z8\") pod \"community-operators-jkflp\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.331152 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:34 crc kubenswrapper[5025]: I1004 11:22:34.816830 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jkflp"] Oct 04 11:22:35 crc kubenswrapper[5025]: I1004 11:22:35.354849 5025 generic.go:334] "Generic (PLEG): container finished" podID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerID="78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962" exitCode=0 Oct 04 11:22:35 crc kubenswrapper[5025]: I1004 11:22:35.354945 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jkflp" event={"ID":"3feb3cdb-d852-4499-8a6c-3d1f420612d7","Type":"ContainerDied","Data":"78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962"} Oct 04 11:22:35 crc kubenswrapper[5025]: I1004 11:22:35.355176 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jkflp" event={"ID":"3feb3cdb-d852-4499-8a6c-3d1f420612d7","Type":"ContainerStarted","Data":"d3c863c25c506cd11358e9b35b6d430fdb849dc1ddcb9b3c7fd2a81fa8551387"} Oct 04 11:22:36 crc kubenswrapper[5025]: I1004 11:22:36.370771 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jkflp" event={"ID":"3feb3cdb-d852-4499-8a6c-3d1f420612d7","Type":"ContainerStarted","Data":"a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1"} Oct 04 11:22:37 crc kubenswrapper[5025]: I1004 11:22:37.408592 5025 generic.go:334] "Generic (PLEG): container finished" podID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerID="a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1" exitCode=0 Oct 04 11:22:37 crc kubenswrapper[5025]: I1004 11:22:37.408776 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jkflp" event={"ID":"3feb3cdb-d852-4499-8a6c-3d1f420612d7","Type":"ContainerDied","Data":"a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1"} Oct 04 11:22:38 crc kubenswrapper[5025]: I1004 11:22:38.428405 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jkflp" event={"ID":"3feb3cdb-d852-4499-8a6c-3d1f420612d7","Type":"ContainerStarted","Data":"a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef"} Oct 04 11:22:38 crc kubenswrapper[5025]: I1004 11:22:38.463453 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jkflp" podStartSLOduration=2.978101034 podStartE2EDuration="5.4634341s" podCreationTimestamp="2025-10-04 11:22:33 +0000 UTC" firstStartedPulling="2025-10-04 11:22:35.357577005 +0000 UTC m=+2883.782543875" lastFinishedPulling="2025-10-04 11:22:37.842910051 +0000 UTC m=+2886.267876941" observedRunningTime="2025-10-04 11:22:38.45411024 +0000 UTC m=+2886.879077150" watchObservedRunningTime="2025-10-04 11:22:38.4634341 +0000 UTC m=+2886.888400970" Oct 04 11:22:44 crc kubenswrapper[5025]: I1004 11:22:44.331337 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:44 crc kubenswrapper[5025]: I1004 11:22:44.331856 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:44 crc kubenswrapper[5025]: I1004 11:22:44.429291 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:44 crc kubenswrapper[5025]: I1004 11:22:44.548832 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:44 crc kubenswrapper[5025]: I1004 11:22:44.686745 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jkflp"] Oct 04 11:22:44 crc kubenswrapper[5025]: I1004 11:22:44.713969 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:22:44 crc kubenswrapper[5025]: I1004 11:22:44.714117 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:22:44 crc kubenswrapper[5025]: I1004 11:22:44.714186 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 11:22:44 crc kubenswrapper[5025]: I1004 11:22:44.715414 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ddf274783bd9ad4cb17c313083966159dc88be7cc9e9b27f856feb0a8e136a6"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:22:44 crc kubenswrapper[5025]: I1004 11:22:44.715546 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://6ddf274783bd9ad4cb17c313083966159dc88be7cc9e9b27f856feb0a8e136a6" gracePeriod=600 Oct 04 11:22:45 crc kubenswrapper[5025]: I1004 11:22:45.515857 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="6ddf274783bd9ad4cb17c313083966159dc88be7cc9e9b27f856feb0a8e136a6" exitCode=0 Oct 04 11:22:45 crc kubenswrapper[5025]: I1004 11:22:45.515933 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"6ddf274783bd9ad4cb17c313083966159dc88be7cc9e9b27f856feb0a8e136a6"} Oct 04 11:22:45 crc kubenswrapper[5025]: I1004 11:22:45.516653 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807"} Oct 04 11:22:45 crc kubenswrapper[5025]: I1004 11:22:45.516687 5025 scope.go:117] "RemoveContainer" containerID="bd81e523fcf731f32cfa9b30982cd33b48ce38feb07cd0bf449ff721af9f7cfa" Oct 04 11:22:46 crc kubenswrapper[5025]: I1004 11:22:46.527199 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jkflp" podUID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerName="registry-server" containerID="cri-o://a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef" gracePeriod=2 Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.012748 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.156948 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-catalog-content\") pod \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.157268 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc7z8\" (UniqueName: \"kubernetes.io/projected/3feb3cdb-d852-4499-8a6c-3d1f420612d7-kube-api-access-jc7z8\") pod \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.157512 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-utilities\") pod \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\" (UID: \"3feb3cdb-d852-4499-8a6c-3d1f420612d7\") " Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.158154 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-utilities" (OuterVolumeSpecName: "utilities") pod "3feb3cdb-d852-4499-8a6c-3d1f420612d7" (UID: "3feb3cdb-d852-4499-8a6c-3d1f420612d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.165583 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3feb3cdb-d852-4499-8a6c-3d1f420612d7-kube-api-access-jc7z8" (OuterVolumeSpecName: "kube-api-access-jc7z8") pod "3feb3cdb-d852-4499-8a6c-3d1f420612d7" (UID: "3feb3cdb-d852-4499-8a6c-3d1f420612d7"). InnerVolumeSpecName "kube-api-access-jc7z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.209265 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3feb3cdb-d852-4499-8a6c-3d1f420612d7" (UID: "3feb3cdb-d852-4499-8a6c-3d1f420612d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.259939 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc7z8\" (UniqueName: \"kubernetes.io/projected/3feb3cdb-d852-4499-8a6c-3d1f420612d7-kube-api-access-jc7z8\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.259994 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.260053 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3feb3cdb-d852-4499-8a6c-3d1f420612d7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.539506 5025 generic.go:334] "Generic (PLEG): container finished" podID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerID="a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef" exitCode=0 Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.539543 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jkflp" event={"ID":"3feb3cdb-d852-4499-8a6c-3d1f420612d7","Type":"ContainerDied","Data":"a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef"} Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.539568 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jkflp" event={"ID":"3feb3cdb-d852-4499-8a6c-3d1f420612d7","Type":"ContainerDied","Data":"d3c863c25c506cd11358e9b35b6d430fdb849dc1ddcb9b3c7fd2a81fa8551387"} Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.539587 5025 scope.go:117] "RemoveContainer" containerID="a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.539703 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jkflp" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.578861 5025 scope.go:117] "RemoveContainer" containerID="a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.581977 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jkflp"] Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.591178 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jkflp"] Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.600320 5025 scope.go:117] "RemoveContainer" containerID="78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.652601 5025 scope.go:117] "RemoveContainer" containerID="a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef" Oct 04 11:22:47 crc kubenswrapper[5025]: E1004 11:22:47.653157 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef\": container with ID starting with a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef not found: ID does not exist" containerID="a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.653212 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef"} err="failed to get container status \"a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef\": rpc error: code = NotFound desc = could not find container \"a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef\": container with ID starting with a70bade93a14edfabc6bcafde68af55abb43b75d74d86d5b91261b4edf24f9ef not found: ID does not exist" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.653246 5025 scope.go:117] "RemoveContainer" containerID="a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1" Oct 04 11:22:47 crc kubenswrapper[5025]: E1004 11:22:47.653633 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1\": container with ID starting with a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1 not found: ID does not exist" containerID="a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.653669 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1"} err="failed to get container status \"a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1\": rpc error: code = NotFound desc = could not find container \"a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1\": container with ID starting with a3a3dd0ee8ef0e5b6219aef09648473acef844f30e70c6522b94880802f97df1 not found: ID does not exist" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.653697 5025 scope.go:117] "RemoveContainer" containerID="78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962" Oct 04 11:22:47 crc kubenswrapper[5025]: E1004 11:22:47.654127 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962\": container with ID starting with 78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962 not found: ID does not exist" containerID="78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962" Oct 04 11:22:47 crc kubenswrapper[5025]: I1004 11:22:47.654155 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962"} err="failed to get container status \"78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962\": rpc error: code = NotFound desc = could not find container \"78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962\": container with ID starting with 78f6de873d0d06862efe6f70e1a9546c14208b4893286e3642bf721061623962 not found: ID does not exist" Oct 04 11:22:48 crc kubenswrapper[5025]: I1004 11:22:48.430333 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" path="/var/lib/kubelet/pods/3feb3cdb-d852-4499-8a6c-3d1f420612d7/volumes" Oct 04 11:25:14 crc kubenswrapper[5025]: I1004 11:25:14.713398 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:25:14 crc kubenswrapper[5025]: I1004 11:25:14.713900 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:25:44 crc kubenswrapper[5025]: I1004 11:25:44.714427 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:25:44 crc kubenswrapper[5025]: I1004 11:25:44.715011 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:26:14 crc kubenswrapper[5025]: I1004 11:26:14.713707 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:26:14 crc kubenswrapper[5025]: I1004 11:26:14.714337 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:26:14 crc kubenswrapper[5025]: I1004 11:26:14.714394 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 11:26:14 crc kubenswrapper[5025]: I1004 11:26:14.715302 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:26:14 crc kubenswrapper[5025]: I1004 11:26:14.715366 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" gracePeriod=600 Oct 04 11:26:14 crc kubenswrapper[5025]: E1004 11:26:14.843274 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:26:15 crc kubenswrapper[5025]: I1004 11:26:15.630419 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" exitCode=0 Oct 04 11:26:15 crc kubenswrapper[5025]: I1004 11:26:15.630491 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807"} Oct 04 11:26:15 crc kubenswrapper[5025]: I1004 11:26:15.630804 5025 scope.go:117] "RemoveContainer" containerID="6ddf274783bd9ad4cb17c313083966159dc88be7cc9e9b27f856feb0a8e136a6" Oct 04 11:26:15 crc kubenswrapper[5025]: I1004 11:26:15.631536 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:26:15 crc kubenswrapper[5025]: E1004 11:26:15.631786 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:26:30 crc kubenswrapper[5025]: I1004 11:26:30.411065 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:26:30 crc kubenswrapper[5025]: E1004 11:26:30.411824 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:26:44 crc kubenswrapper[5025]: I1004 11:26:44.411386 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:26:44 crc kubenswrapper[5025]: E1004 11:26:44.412193 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:26:55 crc kubenswrapper[5025]: I1004 11:26:55.411996 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:26:55 crc kubenswrapper[5025]: E1004 11:26:55.413064 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:27:09 crc kubenswrapper[5025]: I1004 11:27:09.411238 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:27:09 crc kubenswrapper[5025]: E1004 11:27:09.412061 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:27:22 crc kubenswrapper[5025]: I1004 11:27:22.423692 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:27:22 crc kubenswrapper[5025]: E1004 11:27:22.427364 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:27:35 crc kubenswrapper[5025]: I1004 11:27:35.411836 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:27:35 crc kubenswrapper[5025]: E1004 11:27:35.413136 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:27:49 crc kubenswrapper[5025]: I1004 11:27:49.411985 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:27:49 crc kubenswrapper[5025]: E1004 11:27:49.413081 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:28:00 crc kubenswrapper[5025]: I1004 11:28:00.411173 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:28:00 crc kubenswrapper[5025]: E1004 11:28:00.413059 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:28:14 crc kubenswrapper[5025]: I1004 11:28:14.411166 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:28:14 crc kubenswrapper[5025]: E1004 11:28:14.411791 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:28:29 crc kubenswrapper[5025]: I1004 11:28:29.412086 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:28:29 crc kubenswrapper[5025]: E1004 11:28:29.413217 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:28:42 crc kubenswrapper[5025]: I1004 11:28:42.419591 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:28:42 crc kubenswrapper[5025]: E1004 11:28:42.421475 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:28:54 crc kubenswrapper[5025]: I1004 11:28:54.410953 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:28:54 crc kubenswrapper[5025]: E1004 11:28:54.411664 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:29:06 crc kubenswrapper[5025]: I1004 11:29:06.411332 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:29:06 crc kubenswrapper[5025]: E1004 11:29:06.412195 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:29:17 crc kubenswrapper[5025]: I1004 11:29:17.411919 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:29:17 crc kubenswrapper[5025]: E1004 11:29:17.412837 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.222521 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdsj"] Oct 04 11:29:18 crc kubenswrapper[5025]: E1004 11:29:18.222906 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerName="extract-utilities" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.222919 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerName="extract-utilities" Oct 04 11:29:18 crc kubenswrapper[5025]: E1004 11:29:18.222942 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerName="registry-server" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.222950 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerName="registry-server" Oct 04 11:29:18 crc kubenswrapper[5025]: E1004 11:29:18.222958 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerName="extract-content" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.222965 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerName="extract-content" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.223690 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="3feb3cdb-d852-4499-8a6c-3d1f420612d7" containerName="registry-server" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.225067 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.257123 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdsj"] Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.284859 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcvzt\" (UniqueName: \"kubernetes.io/projected/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-kube-api-access-mcvzt\") pod \"redhat-marketplace-bhdsj\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.285129 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-catalog-content\") pod \"redhat-marketplace-bhdsj\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.285203 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-utilities\") pod \"redhat-marketplace-bhdsj\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.387736 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcvzt\" (UniqueName: \"kubernetes.io/projected/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-kube-api-access-mcvzt\") pod \"redhat-marketplace-bhdsj\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.387861 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-catalog-content\") pod \"redhat-marketplace-bhdsj\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.387894 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-utilities\") pod \"redhat-marketplace-bhdsj\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.388402 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-utilities\") pod \"redhat-marketplace-bhdsj\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.388595 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-catalog-content\") pod \"redhat-marketplace-bhdsj\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.418101 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcvzt\" (UniqueName: \"kubernetes.io/projected/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-kube-api-access-mcvzt\") pod \"redhat-marketplace-bhdsj\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:18 crc kubenswrapper[5025]: I1004 11:29:18.548959 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:19 crc kubenswrapper[5025]: I1004 11:29:19.059790 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdsj"] Oct 04 11:29:19 crc kubenswrapper[5025]: I1004 11:29:19.559497 5025 generic.go:334] "Generic (PLEG): container finished" podID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerID="cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449" exitCode=0 Oct 04 11:29:19 crc kubenswrapper[5025]: I1004 11:29:19.559593 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdsj" event={"ID":"6e8d80e2-cf8f-4420-bf12-e1ac52df412f","Type":"ContainerDied","Data":"cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449"} Oct 04 11:29:19 crc kubenswrapper[5025]: I1004 11:29:19.559633 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdsj" event={"ID":"6e8d80e2-cf8f-4420-bf12-e1ac52df412f","Type":"ContainerStarted","Data":"12e38f6f139e622298c246ba4c1c9f155a77dec5a1c8f3d5a8e2b0795dda1388"} Oct 04 11:29:19 crc kubenswrapper[5025]: I1004 11:29:19.562230 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:29:20 crc kubenswrapper[5025]: I1004 11:29:20.572453 5025 generic.go:334] "Generic (PLEG): container finished" podID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerID="2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3" exitCode=0 Oct 04 11:29:20 crc kubenswrapper[5025]: I1004 11:29:20.572733 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdsj" event={"ID":"6e8d80e2-cf8f-4420-bf12-e1ac52df412f","Type":"ContainerDied","Data":"2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3"} Oct 04 11:29:21 crc kubenswrapper[5025]: I1004 11:29:21.584368 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdsj" event={"ID":"6e8d80e2-cf8f-4420-bf12-e1ac52df412f","Type":"ContainerStarted","Data":"76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257"} Oct 04 11:29:21 crc kubenswrapper[5025]: I1004 11:29:21.611780 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bhdsj" podStartSLOduration=2.189358074 podStartE2EDuration="3.611754323s" podCreationTimestamp="2025-10-04 11:29:18 +0000 UTC" firstStartedPulling="2025-10-04 11:29:19.561875507 +0000 UTC m=+3287.986842407" lastFinishedPulling="2025-10-04 11:29:20.984271746 +0000 UTC m=+3289.409238656" observedRunningTime="2025-10-04 11:29:21.605920258 +0000 UTC m=+3290.030887148" watchObservedRunningTime="2025-10-04 11:29:21.611754323 +0000 UTC m=+3290.036721203" Oct 04 11:29:24 crc kubenswrapper[5025]: I1004 11:29:24.825951 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v299m"] Oct 04 11:29:24 crc kubenswrapper[5025]: I1004 11:29:24.829486 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:24 crc kubenswrapper[5025]: I1004 11:29:24.837611 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v299m"] Oct 04 11:29:24 crc kubenswrapper[5025]: I1004 11:29:24.912621 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-catalog-content\") pod \"redhat-operators-v299m\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:24 crc kubenswrapper[5025]: I1004 11:29:24.912863 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-utilities\") pod \"redhat-operators-v299m\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:24 crc kubenswrapper[5025]: I1004 11:29:24.912939 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bqs4\" (UniqueName: \"kubernetes.io/projected/8a4393b7-737b-47fd-8302-cf110fbb4997-kube-api-access-9bqs4\") pod \"redhat-operators-v299m\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:25 crc kubenswrapper[5025]: I1004 11:29:25.015630 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-utilities\") pod \"redhat-operators-v299m\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:25 crc kubenswrapper[5025]: I1004 11:29:25.015971 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bqs4\" (UniqueName: \"kubernetes.io/projected/8a4393b7-737b-47fd-8302-cf110fbb4997-kube-api-access-9bqs4\") pod \"redhat-operators-v299m\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:25 crc kubenswrapper[5025]: I1004 11:29:25.016216 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-utilities\") pod \"redhat-operators-v299m\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:25 crc kubenswrapper[5025]: I1004 11:29:25.016223 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-catalog-content\") pod \"redhat-operators-v299m\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:25 crc kubenswrapper[5025]: I1004 11:29:25.016719 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-catalog-content\") pod \"redhat-operators-v299m\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:25 crc kubenswrapper[5025]: I1004 11:29:25.036889 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bqs4\" (UniqueName: \"kubernetes.io/projected/8a4393b7-737b-47fd-8302-cf110fbb4997-kube-api-access-9bqs4\") pod \"redhat-operators-v299m\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:25 crc kubenswrapper[5025]: I1004 11:29:25.159303 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:25 crc kubenswrapper[5025]: W1004 11:29:25.641707 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a4393b7_737b_47fd_8302_cf110fbb4997.slice/crio-2166b1f865c36dc466e2929d6b812d4895ef518142a369b9d1936da95e991d86 WatchSource:0}: Error finding container 2166b1f865c36dc466e2929d6b812d4895ef518142a369b9d1936da95e991d86: Status 404 returned error can't find the container with id 2166b1f865c36dc466e2929d6b812d4895ef518142a369b9d1936da95e991d86 Oct 04 11:29:25 crc kubenswrapper[5025]: I1004 11:29:25.649114 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v299m"] Oct 04 11:29:26 crc kubenswrapper[5025]: I1004 11:29:26.637611 5025 generic.go:334] "Generic (PLEG): container finished" podID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerID="b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37" exitCode=0 Oct 04 11:29:26 crc kubenswrapper[5025]: I1004 11:29:26.637675 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v299m" event={"ID":"8a4393b7-737b-47fd-8302-cf110fbb4997","Type":"ContainerDied","Data":"b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37"} Oct 04 11:29:26 crc kubenswrapper[5025]: I1004 11:29:26.637713 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v299m" event={"ID":"8a4393b7-737b-47fd-8302-cf110fbb4997","Type":"ContainerStarted","Data":"2166b1f865c36dc466e2929d6b812d4895ef518142a369b9d1936da95e991d86"} Oct 04 11:29:28 crc kubenswrapper[5025]: I1004 11:29:28.412224 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:29:28 crc kubenswrapper[5025]: E1004 11:29:28.415211 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:29:28 crc kubenswrapper[5025]: I1004 11:29:28.549968 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:28 crc kubenswrapper[5025]: I1004 11:29:28.550295 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:28 crc kubenswrapper[5025]: I1004 11:29:28.618064 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:28 crc kubenswrapper[5025]: I1004 11:29:28.665437 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v299m" event={"ID":"8a4393b7-737b-47fd-8302-cf110fbb4997","Type":"ContainerStarted","Data":"da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b"} Oct 04 11:29:28 crc kubenswrapper[5025]: I1004 11:29:28.728562 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:29 crc kubenswrapper[5025]: I1004 11:29:29.676356 5025 generic.go:334] "Generic (PLEG): container finished" podID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerID="da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b" exitCode=0 Oct 04 11:29:29 crc kubenswrapper[5025]: I1004 11:29:29.676444 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v299m" event={"ID":"8a4393b7-737b-47fd-8302-cf110fbb4997","Type":"ContainerDied","Data":"da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b"} Oct 04 11:29:30 crc kubenswrapper[5025]: I1004 11:29:30.694339 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v299m" event={"ID":"8a4393b7-737b-47fd-8302-cf110fbb4997","Type":"ContainerStarted","Data":"413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146"} Oct 04 11:29:30 crc kubenswrapper[5025]: I1004 11:29:30.721988 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v299m" podStartSLOduration=3.245125455 podStartE2EDuration="6.72196691s" podCreationTimestamp="2025-10-04 11:29:24 +0000 UTC" firstStartedPulling="2025-10-04 11:29:26.642216503 +0000 UTC m=+3295.067183423" lastFinishedPulling="2025-10-04 11:29:30.119057998 +0000 UTC m=+3298.544024878" observedRunningTime="2025-10-04 11:29:30.718263326 +0000 UTC m=+3299.143230236" watchObservedRunningTime="2025-10-04 11:29:30.72196691 +0000 UTC m=+3299.146933790" Oct 04 11:29:30 crc kubenswrapper[5025]: I1004 11:29:30.815169 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdsj"] Oct 04 11:29:31 crc kubenswrapper[5025]: I1004 11:29:31.704588 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bhdsj" podUID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerName="registry-server" containerID="cri-o://76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257" gracePeriod=2 Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.260877 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.359322 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-catalog-content\") pod \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.359421 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcvzt\" (UniqueName: \"kubernetes.io/projected/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-kube-api-access-mcvzt\") pod \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.359546 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-utilities\") pod \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\" (UID: \"6e8d80e2-cf8f-4420-bf12-e1ac52df412f\") " Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.360255 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-utilities" (OuterVolumeSpecName: "utilities") pod "6e8d80e2-cf8f-4420-bf12-e1ac52df412f" (UID: "6e8d80e2-cf8f-4420-bf12-e1ac52df412f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.364795 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-kube-api-access-mcvzt" (OuterVolumeSpecName: "kube-api-access-mcvzt") pod "6e8d80e2-cf8f-4420-bf12-e1ac52df412f" (UID: "6e8d80e2-cf8f-4420-bf12-e1ac52df412f"). InnerVolumeSpecName "kube-api-access-mcvzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.370964 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6e8d80e2-cf8f-4420-bf12-e1ac52df412f" (UID: "6e8d80e2-cf8f-4420-bf12-e1ac52df412f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.461999 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.462039 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.462050 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcvzt\" (UniqueName: \"kubernetes.io/projected/6e8d80e2-cf8f-4420-bf12-e1ac52df412f-kube-api-access-mcvzt\") on node \"crc\" DevicePath \"\"" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.718309 5025 generic.go:334] "Generic (PLEG): container finished" podID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerID="76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257" exitCode=0 Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.718354 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhdsj" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.718357 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdsj" event={"ID":"6e8d80e2-cf8f-4420-bf12-e1ac52df412f","Type":"ContainerDied","Data":"76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257"} Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.718570 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhdsj" event={"ID":"6e8d80e2-cf8f-4420-bf12-e1ac52df412f","Type":"ContainerDied","Data":"12e38f6f139e622298c246ba4c1c9f155a77dec5a1c8f3d5a8e2b0795dda1388"} Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.718637 5025 scope.go:117] "RemoveContainer" containerID="76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.760209 5025 scope.go:117] "RemoveContainer" containerID="2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.775818 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdsj"] Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.791653 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhdsj"] Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.795047 5025 scope.go:117] "RemoveContainer" containerID="cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.863283 5025 scope.go:117] "RemoveContainer" containerID="76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257" Oct 04 11:29:32 crc kubenswrapper[5025]: E1004 11:29:32.863844 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257\": container with ID starting with 76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257 not found: ID does not exist" containerID="76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.863892 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257"} err="failed to get container status \"76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257\": rpc error: code = NotFound desc = could not find container \"76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257\": container with ID starting with 76b3ec8fa52609e2e6c9dce47fecb7fe6805110c2d0a1beae6c664856f96a257 not found: ID does not exist" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.863928 5025 scope.go:117] "RemoveContainer" containerID="2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3" Oct 04 11:29:32 crc kubenswrapper[5025]: E1004 11:29:32.864274 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3\": container with ID starting with 2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3 not found: ID does not exist" containerID="2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.864315 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3"} err="failed to get container status \"2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3\": rpc error: code = NotFound desc = could not find container \"2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3\": container with ID starting with 2a0f94986233e84b6a756af7eb7e579bd96c8bb467b041c0b900785d7612dfd3 not found: ID does not exist" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.864339 5025 scope.go:117] "RemoveContainer" containerID="cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449" Oct 04 11:29:32 crc kubenswrapper[5025]: E1004 11:29:32.864721 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449\": container with ID starting with cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449 not found: ID does not exist" containerID="cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449" Oct 04 11:29:32 crc kubenswrapper[5025]: I1004 11:29:32.864755 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449"} err="failed to get container status \"cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449\": rpc error: code = NotFound desc = could not find container \"cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449\": container with ID starting with cd491a61c203968b41e3d70e69692498f65c2d42c4b446ec8ea65d4882917449 not found: ID does not exist" Oct 04 11:29:34 crc kubenswrapper[5025]: I1004 11:29:34.426776 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" path="/var/lib/kubelet/pods/6e8d80e2-cf8f-4420-bf12-e1ac52df412f/volumes" Oct 04 11:29:35 crc kubenswrapper[5025]: I1004 11:29:35.160428 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:35 crc kubenswrapper[5025]: I1004 11:29:35.160784 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:35 crc kubenswrapper[5025]: I1004 11:29:35.216364 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:35 crc kubenswrapper[5025]: E1004 11:29:35.600544 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice/crio-12e38f6f139e622298c246ba4c1c9f155a77dec5a1c8f3d5a8e2b0795dda1388\": RecentStats: unable to find data in memory cache]" Oct 04 11:29:35 crc kubenswrapper[5025]: I1004 11:29:35.821201 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:37 crc kubenswrapper[5025]: I1004 11:29:37.012676 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v299m"] Oct 04 11:29:37 crc kubenswrapper[5025]: I1004 11:29:37.776269 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v299m" podUID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerName="registry-server" containerID="cri-o://413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146" gracePeriod=2 Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.328159 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.494242 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bqs4\" (UniqueName: \"kubernetes.io/projected/8a4393b7-737b-47fd-8302-cf110fbb4997-kube-api-access-9bqs4\") pod \"8a4393b7-737b-47fd-8302-cf110fbb4997\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.494458 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-catalog-content\") pod \"8a4393b7-737b-47fd-8302-cf110fbb4997\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.494547 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-utilities\") pod \"8a4393b7-737b-47fd-8302-cf110fbb4997\" (UID: \"8a4393b7-737b-47fd-8302-cf110fbb4997\") " Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.495324 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-utilities" (OuterVolumeSpecName: "utilities") pod "8a4393b7-737b-47fd-8302-cf110fbb4997" (UID: "8a4393b7-737b-47fd-8302-cf110fbb4997"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.501595 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a4393b7-737b-47fd-8302-cf110fbb4997-kube-api-access-9bqs4" (OuterVolumeSpecName: "kube-api-access-9bqs4") pod "8a4393b7-737b-47fd-8302-cf110fbb4997" (UID: "8a4393b7-737b-47fd-8302-cf110fbb4997"). InnerVolumeSpecName "kube-api-access-9bqs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.600247 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.600293 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bqs4\" (UniqueName: \"kubernetes.io/projected/8a4393b7-737b-47fd-8302-cf110fbb4997-kube-api-access-9bqs4\") on node \"crc\" DevicePath \"\"" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.617945 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a4393b7-737b-47fd-8302-cf110fbb4997" (UID: "8a4393b7-737b-47fd-8302-cf110fbb4997"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.702505 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a4393b7-737b-47fd-8302-cf110fbb4997-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.792798 5025 generic.go:334] "Generic (PLEG): container finished" podID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerID="413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146" exitCode=0 Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.792852 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v299m" event={"ID":"8a4393b7-737b-47fd-8302-cf110fbb4997","Type":"ContainerDied","Data":"413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146"} Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.792890 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v299m" event={"ID":"8a4393b7-737b-47fd-8302-cf110fbb4997","Type":"ContainerDied","Data":"2166b1f865c36dc466e2929d6b812d4895ef518142a369b9d1936da95e991d86"} Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.792917 5025 scope.go:117] "RemoveContainer" containerID="413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.793138 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v299m" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.826571 5025 scope.go:117] "RemoveContainer" containerID="da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.852841 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v299m"] Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.858598 5025 scope.go:117] "RemoveContainer" containerID="b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.863661 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v299m"] Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.891524 5025 scope.go:117] "RemoveContainer" containerID="413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146" Oct 04 11:29:38 crc kubenswrapper[5025]: E1004 11:29:38.891898 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146\": container with ID starting with 413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146 not found: ID does not exist" containerID="413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.891929 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146"} err="failed to get container status \"413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146\": rpc error: code = NotFound desc = could not find container \"413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146\": container with ID starting with 413408a264a93cffcb5f79307262ce95a67f69adb291893275472d0262c42146 not found: ID does not exist" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.891951 5025 scope.go:117] "RemoveContainer" containerID="da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b" Oct 04 11:29:38 crc kubenswrapper[5025]: E1004 11:29:38.893260 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b\": container with ID starting with da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b not found: ID does not exist" containerID="da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.893330 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b"} err="failed to get container status \"da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b\": rpc error: code = NotFound desc = could not find container \"da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b\": container with ID starting with da952a0cf98a021376a02da1a8b4b104a824e49fd4c7b7bd580658416100bd6b not found: ID does not exist" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.893359 5025 scope.go:117] "RemoveContainer" containerID="b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37" Oct 04 11:29:38 crc kubenswrapper[5025]: E1004 11:29:38.893653 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37\": container with ID starting with b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37 not found: ID does not exist" containerID="b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37" Oct 04 11:29:38 crc kubenswrapper[5025]: I1004 11:29:38.893680 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37"} err="failed to get container status \"b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37\": rpc error: code = NotFound desc = could not find container \"b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37\": container with ID starting with b80706f2351b58676579849c9fbaee92fe23a6a7a263c5b653b85655c1e9ca37 not found: ID does not exist" Oct 04 11:29:39 crc kubenswrapper[5025]: I1004 11:29:39.412338 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:29:39 crc kubenswrapper[5025]: E1004 11:29:39.412591 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:29:40 crc kubenswrapper[5025]: I1004 11:29:40.431552 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a4393b7-737b-47fd-8302-cf110fbb4997" path="/var/lib/kubelet/pods/8a4393b7-737b-47fd-8302-cf110fbb4997/volumes" Oct 04 11:29:45 crc kubenswrapper[5025]: E1004 11:29:45.892143 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice/crio-12e38f6f139e622298c246ba4c1c9f155a77dec5a1c8f3d5a8e2b0795dda1388\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice\": RecentStats: unable to find data in memory cache]" Oct 04 11:29:50 crc kubenswrapper[5025]: I1004 11:29:50.411346 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:29:50 crc kubenswrapper[5025]: E1004 11:29:50.412383 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:29:56 crc kubenswrapper[5025]: E1004 11:29:56.181795 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice/crio-12e38f6f139e622298c246ba4c1c9f155a77dec5a1c8f3d5a8e2b0795dda1388\": RecentStats: unable to find data in memory cache]" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.187654 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g"] Oct 04 11:30:00 crc kubenswrapper[5025]: E1004 11:30:00.190070 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerName="extract-content" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.190181 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerName="extract-content" Oct 04 11:30:00 crc kubenswrapper[5025]: E1004 11:30:00.190271 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerName="extract-content" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.190348 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerName="extract-content" Oct 04 11:30:00 crc kubenswrapper[5025]: E1004 11:30:00.190441 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerName="extract-utilities" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.190515 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerName="extract-utilities" Oct 04 11:30:00 crc kubenswrapper[5025]: E1004 11:30:00.190602 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerName="extract-utilities" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.190677 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerName="extract-utilities" Oct 04 11:30:00 crc kubenswrapper[5025]: E1004 11:30:00.190770 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerName="registry-server" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.190854 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerName="registry-server" Oct 04 11:30:00 crc kubenswrapper[5025]: E1004 11:30:00.190963 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerName="registry-server" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.191074 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerName="registry-server" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.191460 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e8d80e2-cf8f-4420-bf12-e1ac52df412f" containerName="registry-server" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.191572 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a4393b7-737b-47fd-8302-cf110fbb4997" containerName="registry-server" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.195034 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.198148 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.198187 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.210991 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g"] Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.349176 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsr8p\" (UniqueName: \"kubernetes.io/projected/791b3ce3-d1c6-409d-a107-7cd44a5c9271-kube-api-access-jsr8p\") pod \"collect-profiles-29326290-84r4g\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.349244 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/791b3ce3-d1c6-409d-a107-7cd44a5c9271-config-volume\") pod \"collect-profiles-29326290-84r4g\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.349478 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/791b3ce3-d1c6-409d-a107-7cd44a5c9271-secret-volume\") pod \"collect-profiles-29326290-84r4g\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.451516 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/791b3ce3-d1c6-409d-a107-7cd44a5c9271-secret-volume\") pod \"collect-profiles-29326290-84r4g\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.451658 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsr8p\" (UniqueName: \"kubernetes.io/projected/791b3ce3-d1c6-409d-a107-7cd44a5c9271-kube-api-access-jsr8p\") pod \"collect-profiles-29326290-84r4g\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.451690 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/791b3ce3-d1c6-409d-a107-7cd44a5c9271-config-volume\") pod \"collect-profiles-29326290-84r4g\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.452604 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/791b3ce3-d1c6-409d-a107-7cd44a5c9271-config-volume\") pod \"collect-profiles-29326290-84r4g\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.462869 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/791b3ce3-d1c6-409d-a107-7cd44a5c9271-secret-volume\") pod \"collect-profiles-29326290-84r4g\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.473342 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsr8p\" (UniqueName: \"kubernetes.io/projected/791b3ce3-d1c6-409d-a107-7cd44a5c9271-kube-api-access-jsr8p\") pod \"collect-profiles-29326290-84r4g\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.523511 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:00 crc kubenswrapper[5025]: I1004 11:30:00.978496 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g"] Oct 04 11:30:00 crc kubenswrapper[5025]: W1004 11:30:00.987221 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod791b3ce3_d1c6_409d_a107_7cd44a5c9271.slice/crio-e902d28dd9a7a3bafd5d11ee9e6579cceb0b860375d101372ffa43e43236b09f WatchSource:0}: Error finding container e902d28dd9a7a3bafd5d11ee9e6579cceb0b860375d101372ffa43e43236b09f: Status 404 returned error can't find the container with id e902d28dd9a7a3bafd5d11ee9e6579cceb0b860375d101372ffa43e43236b09f Oct 04 11:30:01 crc kubenswrapper[5025]: I1004 11:30:01.024715 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" event={"ID":"791b3ce3-d1c6-409d-a107-7cd44a5c9271","Type":"ContainerStarted","Data":"e902d28dd9a7a3bafd5d11ee9e6579cceb0b860375d101372ffa43e43236b09f"} Oct 04 11:30:02 crc kubenswrapper[5025]: I1004 11:30:02.038974 5025 generic.go:334] "Generic (PLEG): container finished" podID="791b3ce3-d1c6-409d-a107-7cd44a5c9271" containerID="0564e1d60b0d76303a399e2a5d2479adb91e29d723aec8ae398df3e5e53f176f" exitCode=0 Oct 04 11:30:02 crc kubenswrapper[5025]: I1004 11:30:02.039093 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" event={"ID":"791b3ce3-d1c6-409d-a107-7cd44a5c9271","Type":"ContainerDied","Data":"0564e1d60b0d76303a399e2a5d2479adb91e29d723aec8ae398df3e5e53f176f"} Oct 04 11:30:03 crc kubenswrapper[5025]: I1004 11:30:03.434092 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:03 crc kubenswrapper[5025]: I1004 11:30:03.508650 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/791b3ce3-d1c6-409d-a107-7cd44a5c9271-config-volume\") pod \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " Oct 04 11:30:03 crc kubenswrapper[5025]: I1004 11:30:03.509369 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/791b3ce3-d1c6-409d-a107-7cd44a5c9271-secret-volume\") pod \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " Oct 04 11:30:03 crc kubenswrapper[5025]: I1004 11:30:03.509478 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsr8p\" (UniqueName: \"kubernetes.io/projected/791b3ce3-d1c6-409d-a107-7cd44a5c9271-kube-api-access-jsr8p\") pod \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\" (UID: \"791b3ce3-d1c6-409d-a107-7cd44a5c9271\") " Oct 04 11:30:03 crc kubenswrapper[5025]: I1004 11:30:03.510620 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/791b3ce3-d1c6-409d-a107-7cd44a5c9271-config-volume" (OuterVolumeSpecName: "config-volume") pod "791b3ce3-d1c6-409d-a107-7cd44a5c9271" (UID: "791b3ce3-d1c6-409d-a107-7cd44a5c9271"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:30:03 crc kubenswrapper[5025]: I1004 11:30:03.519449 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/791b3ce3-d1c6-409d-a107-7cd44a5c9271-kube-api-access-jsr8p" (OuterVolumeSpecName: "kube-api-access-jsr8p") pod "791b3ce3-d1c6-409d-a107-7cd44a5c9271" (UID: "791b3ce3-d1c6-409d-a107-7cd44a5c9271"). InnerVolumeSpecName "kube-api-access-jsr8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:30:03 crc kubenswrapper[5025]: I1004 11:30:03.519612 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/791b3ce3-d1c6-409d-a107-7cd44a5c9271-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "791b3ce3-d1c6-409d-a107-7cd44a5c9271" (UID: "791b3ce3-d1c6-409d-a107-7cd44a5c9271"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:30:03 crc kubenswrapper[5025]: I1004 11:30:03.611865 5025 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/791b3ce3-d1c6-409d-a107-7cd44a5c9271-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:03 crc kubenswrapper[5025]: I1004 11:30:03.611899 5025 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/791b3ce3-d1c6-409d-a107-7cd44a5c9271-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:03 crc kubenswrapper[5025]: I1004 11:30:03.611911 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsr8p\" (UniqueName: \"kubernetes.io/projected/791b3ce3-d1c6-409d-a107-7cd44a5c9271-kube-api-access-jsr8p\") on node \"crc\" DevicePath \"\"" Oct 04 11:30:04 crc kubenswrapper[5025]: I1004 11:30:04.057378 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" event={"ID":"791b3ce3-d1c6-409d-a107-7cd44a5c9271","Type":"ContainerDied","Data":"e902d28dd9a7a3bafd5d11ee9e6579cceb0b860375d101372ffa43e43236b09f"} Oct 04 11:30:04 crc kubenswrapper[5025]: I1004 11:30:04.057690 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e902d28dd9a7a3bafd5d11ee9e6579cceb0b860375d101372ffa43e43236b09f" Oct 04 11:30:04 crc kubenswrapper[5025]: I1004 11:30:04.057425 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326290-84r4g" Oct 04 11:30:04 crc kubenswrapper[5025]: I1004 11:30:04.511710 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk"] Oct 04 11:30:04 crc kubenswrapper[5025]: I1004 11:30:04.521746 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326245-glndk"] Oct 04 11:30:05 crc kubenswrapper[5025]: I1004 11:30:05.411854 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:30:05 crc kubenswrapper[5025]: E1004 11:30:05.412380 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:30:06 crc kubenswrapper[5025]: I1004 11:30:06.432541 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8" path="/var/lib/kubelet/pods/96fc8ecd-3f6e-4bd6-ad7c-6ca2145883a8/volumes" Oct 04 11:30:06 crc kubenswrapper[5025]: E1004 11:30:06.439046 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice/crio-12e38f6f139e622298c246ba4c1c9f155a77dec5a1c8f3d5a8e2b0795dda1388\": RecentStats: unable to find data in memory cache]" Oct 04 11:30:16 crc kubenswrapper[5025]: E1004 11:30:16.701819 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice/crio-12e38f6f139e622298c246ba4c1c9f155a77dec5a1c8f3d5a8e2b0795dda1388\": RecentStats: unable to find data in memory cache]" Oct 04 11:30:19 crc kubenswrapper[5025]: I1004 11:30:19.413336 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:30:19 crc kubenswrapper[5025]: E1004 11:30:19.414136 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:30:26 crc kubenswrapper[5025]: E1004 11:30:26.987794 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e8d80e2_cf8f_4420_bf12_e1ac52df412f.slice/crio-12e38f6f139e622298c246ba4c1c9f155a77dec5a1c8f3d5a8e2b0795dda1388\": RecentStats: unable to find data in memory cache]" Oct 04 11:30:32 crc kubenswrapper[5025]: E1004 11:30:32.446220 5025 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/b95889ad96241f13aae475f69cf4fdc3c349a5191dfbb0303fc9bcec30bd11ea/diff" to get inode usage: stat /var/lib/containers/storage/overlay/b95889ad96241f13aae475f69cf4fdc3c349a5191dfbb0303fc9bcec30bd11ea/diff: no such file or directory, extraDiskErr: Oct 04 11:30:34 crc kubenswrapper[5025]: I1004 11:30:34.411451 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:30:34 crc kubenswrapper[5025]: E1004 11:30:34.412427 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:30:48 crc kubenswrapper[5025]: I1004 11:30:48.412112 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:30:48 crc kubenswrapper[5025]: E1004 11:30:48.413286 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:30:50 crc kubenswrapper[5025]: I1004 11:30:50.180114 5025 scope.go:117] "RemoveContainer" containerID="11aac83d485b7ae8776577868850b293b8daa24f8cb8b1da3f1fc992e40c2afb" Oct 04 11:31:00 crc kubenswrapper[5025]: I1004 11:31:00.411084 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:31:00 crc kubenswrapper[5025]: E1004 11:31:00.411785 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:31:11 crc kubenswrapper[5025]: I1004 11:31:11.411660 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:31:11 crc kubenswrapper[5025]: E1004 11:31:11.412489 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:31:22 crc kubenswrapper[5025]: I1004 11:31:22.423214 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:31:22 crc kubenswrapper[5025]: I1004 11:31:22.868393 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"f262faf2a678238fd476697f624cd66122884cd0f3538b657c6e0e36f229e229"} Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.215937 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6xbwk"] Oct 04 11:31:52 crc kubenswrapper[5025]: E1004 11:31:52.218660 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791b3ce3-d1c6-409d-a107-7cd44a5c9271" containerName="collect-profiles" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.218819 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="791b3ce3-d1c6-409d-a107-7cd44a5c9271" containerName="collect-profiles" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.219288 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="791b3ce3-d1c6-409d-a107-7cd44a5c9271" containerName="collect-profiles" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.227683 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xbwk"] Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.229155 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.360510 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-utilities\") pod \"certified-operators-6xbwk\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.360605 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-catalog-content\") pod \"certified-operators-6xbwk\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.360766 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j6q9\" (UniqueName: \"kubernetes.io/projected/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-kube-api-access-5j6q9\") pod \"certified-operators-6xbwk\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.462601 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j6q9\" (UniqueName: \"kubernetes.io/projected/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-kube-api-access-5j6q9\") pod \"certified-operators-6xbwk\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.462708 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-utilities\") pod \"certified-operators-6xbwk\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.462735 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-catalog-content\") pod \"certified-operators-6xbwk\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.463250 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-catalog-content\") pod \"certified-operators-6xbwk\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.463259 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-utilities\") pod \"certified-operators-6xbwk\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.486613 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j6q9\" (UniqueName: \"kubernetes.io/projected/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-kube-api-access-5j6q9\") pod \"certified-operators-6xbwk\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:52 crc kubenswrapper[5025]: I1004 11:31:52.559401 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:31:53 crc kubenswrapper[5025]: I1004 11:31:53.054573 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xbwk"] Oct 04 11:31:53 crc kubenswrapper[5025]: I1004 11:31:53.177227 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbwk" event={"ID":"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5","Type":"ContainerStarted","Data":"7d0ef01288ce48747c1640214aa649845a33bdefeb92c8649473b58174c79760"} Oct 04 11:31:54 crc kubenswrapper[5025]: I1004 11:31:54.186701 5025 generic.go:334] "Generic (PLEG): container finished" podID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerID="4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448" exitCode=0 Oct 04 11:31:54 crc kubenswrapper[5025]: I1004 11:31:54.186746 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbwk" event={"ID":"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5","Type":"ContainerDied","Data":"4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448"} Oct 04 11:31:55 crc kubenswrapper[5025]: I1004 11:31:55.199595 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbwk" event={"ID":"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5","Type":"ContainerStarted","Data":"e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40"} Oct 04 11:31:56 crc kubenswrapper[5025]: I1004 11:31:56.211744 5025 generic.go:334] "Generic (PLEG): container finished" podID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerID="e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40" exitCode=0 Oct 04 11:31:56 crc kubenswrapper[5025]: I1004 11:31:56.211847 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbwk" event={"ID":"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5","Type":"ContainerDied","Data":"e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40"} Oct 04 11:31:57 crc kubenswrapper[5025]: I1004 11:31:57.230512 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbwk" event={"ID":"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5","Type":"ContainerStarted","Data":"addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233"} Oct 04 11:31:57 crc kubenswrapper[5025]: I1004 11:31:57.251538 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6xbwk" podStartSLOduration=2.804854911 podStartE2EDuration="5.251516331s" podCreationTimestamp="2025-10-04 11:31:52 +0000 UTC" firstStartedPulling="2025-10-04 11:31:54.188457528 +0000 UTC m=+3442.613424408" lastFinishedPulling="2025-10-04 11:31:56.635118938 +0000 UTC m=+3445.060085828" observedRunningTime="2025-10-04 11:31:57.24581701 +0000 UTC m=+3445.670783890" watchObservedRunningTime="2025-10-04 11:31:57.251516331 +0000 UTC m=+3445.676483231" Oct 04 11:32:02 crc kubenswrapper[5025]: I1004 11:32:02.559852 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:32:02 crc kubenswrapper[5025]: I1004 11:32:02.560727 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:32:02 crc kubenswrapper[5025]: I1004 11:32:02.627918 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:32:03 crc kubenswrapper[5025]: I1004 11:32:03.357055 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:32:03 crc kubenswrapper[5025]: I1004 11:32:03.411977 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xbwk"] Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.300605 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6xbwk" podUID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerName="registry-server" containerID="cri-o://addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233" gracePeriod=2 Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.739809 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.828667 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-utilities\") pod \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.828856 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j6q9\" (UniqueName: \"kubernetes.io/projected/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-kube-api-access-5j6q9\") pod \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.828895 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-catalog-content\") pod \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\" (UID: \"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5\") " Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.829583 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-utilities" (OuterVolumeSpecName: "utilities") pod "24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" (UID: "24ffb05b-9b61-43b3-b7a7-26aa42eec6b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.841602 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-kube-api-access-5j6q9" (OuterVolumeSpecName: "kube-api-access-5j6q9") pod "24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" (UID: "24ffb05b-9b61-43b3-b7a7-26aa42eec6b5"). InnerVolumeSpecName "kube-api-access-5j6q9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.872546 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" (UID: "24ffb05b-9b61-43b3-b7a7-26aa42eec6b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.931371 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j6q9\" (UniqueName: \"kubernetes.io/projected/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-kube-api-access-5j6q9\") on node \"crc\" DevicePath \"\"" Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.931630 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:32:05 crc kubenswrapper[5025]: I1004 11:32:05.931644 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.311975 5025 generic.go:334] "Generic (PLEG): container finished" podID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerID="addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233" exitCode=0 Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.312067 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xbwk" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.312084 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbwk" event={"ID":"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5","Type":"ContainerDied","Data":"addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233"} Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.312123 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xbwk" event={"ID":"24ffb05b-9b61-43b3-b7a7-26aa42eec6b5","Type":"ContainerDied","Data":"7d0ef01288ce48747c1640214aa649845a33bdefeb92c8649473b58174c79760"} Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.312146 5025 scope.go:117] "RemoveContainer" containerID="addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.334676 5025 scope.go:117] "RemoveContainer" containerID="e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.353528 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xbwk"] Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.360130 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6xbwk"] Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.371244 5025 scope.go:117] "RemoveContainer" containerID="4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.406224 5025 scope.go:117] "RemoveContainer" containerID="addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233" Oct 04 11:32:06 crc kubenswrapper[5025]: E1004 11:32:06.406703 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233\": container with ID starting with addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233 not found: ID does not exist" containerID="addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.406735 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233"} err="failed to get container status \"addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233\": rpc error: code = NotFound desc = could not find container \"addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233\": container with ID starting with addb530f82d5e2f2e198d054f02217a83207ac731cd8e04f5a75cd2cc0712233 not found: ID does not exist" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.406758 5025 scope.go:117] "RemoveContainer" containerID="e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40" Oct 04 11:32:06 crc kubenswrapper[5025]: E1004 11:32:06.407134 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40\": container with ID starting with e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40 not found: ID does not exist" containerID="e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.407159 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40"} err="failed to get container status \"e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40\": rpc error: code = NotFound desc = could not find container \"e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40\": container with ID starting with e9c70128ec60e4092790896da00ab2a8b2334a93409507bb0588f6ca28af5b40 not found: ID does not exist" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.407176 5025 scope.go:117] "RemoveContainer" containerID="4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448" Oct 04 11:32:06 crc kubenswrapper[5025]: E1004 11:32:06.407482 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448\": container with ID starting with 4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448 not found: ID does not exist" containerID="4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.407509 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448"} err="failed to get container status \"4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448\": rpc error: code = NotFound desc = could not find container \"4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448\": container with ID starting with 4c16feefe8d38807ec16a02317a6fb33a85bdb14643999296103601ce0172448 not found: ID does not exist" Oct 04 11:32:06 crc kubenswrapper[5025]: I1004 11:32:06.422718 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" path="/var/lib/kubelet/pods/24ffb05b-9b61-43b3-b7a7-26aa42eec6b5/volumes" Oct 04 11:33:12 crc kubenswrapper[5025]: I1004 11:33:12.002555 5025 generic.go:334] "Generic (PLEG): container finished" podID="5c783e4c-998b-44ab-956a-de26d5568f90" containerID="0abe21bb6337ddcfb19fd7f1b1d8a78ef17f4cf75fcd6542f3e7ecb835320382" exitCode=0 Oct 04 11:33:12 crc kubenswrapper[5025]: I1004 11:33:12.002593 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5c783e4c-998b-44ab-956a-de26d5568f90","Type":"ContainerDied","Data":"0abe21bb6337ddcfb19fd7f1b1d8a78ef17f4cf75fcd6542f3e7ecb835320382"} Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.401348 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.478098 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config\") pod \"5c783e4c-998b-44ab-956a-de26d5568f90\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.478208 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-temporary\") pod \"5c783e4c-998b-44ab-956a-de26d5568f90\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.478279 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzzgs\" (UniqueName: \"kubernetes.io/projected/5c783e4c-998b-44ab-956a-de26d5568f90-kube-api-access-hzzgs\") pod \"5c783e4c-998b-44ab-956a-de26d5568f90\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.478309 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ca-certs\") pod \"5c783e4c-998b-44ab-956a-de26d5568f90\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.478337 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"5c783e4c-998b-44ab-956a-de26d5568f90\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.478367 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-config-data\") pod \"5c783e4c-998b-44ab-956a-de26d5568f90\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.478407 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-workdir\") pod \"5c783e4c-998b-44ab-956a-de26d5568f90\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.478493 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config-secret\") pod \"5c783e4c-998b-44ab-956a-de26d5568f90\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.478573 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ssh-key\") pod \"5c783e4c-998b-44ab-956a-de26d5568f90\" (UID: \"5c783e4c-998b-44ab-956a-de26d5568f90\") " Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.483266 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-config-data" (OuterVolumeSpecName: "config-data") pod "5c783e4c-998b-44ab-956a-de26d5568f90" (UID: "5c783e4c-998b-44ab-956a-de26d5568f90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.485060 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "5c783e4c-998b-44ab-956a-de26d5568f90" (UID: "5c783e4c-998b-44ab-956a-de26d5568f90"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.486359 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "5c783e4c-998b-44ab-956a-de26d5568f90" (UID: "5c783e4c-998b-44ab-956a-de26d5568f90"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.487401 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "5c783e4c-998b-44ab-956a-de26d5568f90" (UID: "5c783e4c-998b-44ab-956a-de26d5568f90"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.489712 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c783e4c-998b-44ab-956a-de26d5568f90-kube-api-access-hzzgs" (OuterVolumeSpecName: "kube-api-access-hzzgs") pod "5c783e4c-998b-44ab-956a-de26d5568f90" (UID: "5c783e4c-998b-44ab-956a-de26d5568f90"). InnerVolumeSpecName "kube-api-access-hzzgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.509332 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "5c783e4c-998b-44ab-956a-de26d5568f90" (UID: "5c783e4c-998b-44ab-956a-de26d5568f90"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.516741 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5c783e4c-998b-44ab-956a-de26d5568f90" (UID: "5c783e4c-998b-44ab-956a-de26d5568f90"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.523535 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5c783e4c-998b-44ab-956a-de26d5568f90" (UID: "5c783e4c-998b-44ab-956a-de26d5568f90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.553747 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5c783e4c-998b-44ab-956a-de26d5568f90" (UID: "5c783e4c-998b-44ab-956a-de26d5568f90"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.581836 5025 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.581875 5025 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.581889 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzzgs\" (UniqueName: \"kubernetes.io/projected/5c783e4c-998b-44ab-956a-de26d5568f90-kube-api-access-hzzgs\") on node \"crc\" DevicePath \"\"" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.581903 5025 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.581940 5025 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.581953 5025 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c783e4c-998b-44ab-956a-de26d5568f90-config-data\") on node \"crc\" DevicePath \"\"" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.581964 5025 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5c783e4c-998b-44ab-956a-de26d5568f90-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.581975 5025 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.581985 5025 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5c783e4c-998b-44ab-956a-de26d5568f90-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.602051 5025 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 04 11:33:13 crc kubenswrapper[5025]: I1004 11:33:13.687006 5025 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 04 11:33:14 crc kubenswrapper[5025]: I1004 11:33:14.026562 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5c783e4c-998b-44ab-956a-de26d5568f90","Type":"ContainerDied","Data":"1477d226cc55bb59c5d7a84a91872a12f4a9e479593e9257c60379cd7ba4c507"} Oct 04 11:33:14 crc kubenswrapper[5025]: I1004 11:33:14.026601 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1477d226cc55bb59c5d7a84a91872a12f4a9e479593e9257c60379cd7ba4c507" Oct 04 11:33:14 crc kubenswrapper[5025]: I1004 11:33:14.026620 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.407443 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 11:33:19 crc kubenswrapper[5025]: E1004 11:33:19.408440 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerName="extract-content" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.408461 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerName="extract-content" Oct 04 11:33:19 crc kubenswrapper[5025]: E1004 11:33:19.408490 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c783e4c-998b-44ab-956a-de26d5568f90" containerName="tempest-tests-tempest-tests-runner" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.408498 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c783e4c-998b-44ab-956a-de26d5568f90" containerName="tempest-tests-tempest-tests-runner" Oct 04 11:33:19 crc kubenswrapper[5025]: E1004 11:33:19.408520 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerName="extract-utilities" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.408529 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerName="extract-utilities" Oct 04 11:33:19 crc kubenswrapper[5025]: E1004 11:33:19.408557 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerName="registry-server" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.408564 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerName="registry-server" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.408774 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c783e4c-998b-44ab-956a-de26d5568f90" containerName="tempest-tests-tempest-tests-runner" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.408794 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ffb05b-9b61-43b3-b7a7-26aa42eec6b5" containerName="registry-server" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.409513 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.411684 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-bdqc5" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.415597 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.502213 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea4d0d6d-6d8b-45c3-abbb-d69654bc2622\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.502452 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvvfg\" (UniqueName: \"kubernetes.io/projected/ea4d0d6d-6d8b-45c3-abbb-d69654bc2622-kube-api-access-hvvfg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea4d0d6d-6d8b-45c3-abbb-d69654bc2622\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.604754 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvvfg\" (UniqueName: \"kubernetes.io/projected/ea4d0d6d-6d8b-45c3-abbb-d69654bc2622-kube-api-access-hvvfg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea4d0d6d-6d8b-45c3-abbb-d69654bc2622\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.604862 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea4d0d6d-6d8b-45c3-abbb-d69654bc2622\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.605250 5025 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea4d0d6d-6d8b-45c3-abbb-d69654bc2622\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.625934 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvvfg\" (UniqueName: \"kubernetes.io/projected/ea4d0d6d-6d8b-45c3-abbb-d69654bc2622-kube-api-access-hvvfg\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea4d0d6d-6d8b-45c3-abbb-d69654bc2622\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.636924 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"ea4d0d6d-6d8b-45c3-abbb-d69654bc2622\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:33:19 crc kubenswrapper[5025]: I1004 11:33:19.734836 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 04 11:33:20 crc kubenswrapper[5025]: I1004 11:33:20.202332 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 04 11:33:21 crc kubenswrapper[5025]: I1004 11:33:21.109372 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"ea4d0d6d-6d8b-45c3-abbb-d69654bc2622","Type":"ContainerStarted","Data":"edd9c25ed7033b37910f493c50fe282096b46d8635add7e2d84fc067e86b55c8"} Oct 04 11:33:22 crc kubenswrapper[5025]: I1004 11:33:22.118836 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"ea4d0d6d-6d8b-45c3-abbb-d69654bc2622","Type":"ContainerStarted","Data":"a9c52cb91ab6225404c5d0b2093714a3d849e36a9bf50daaf9eb3b8913ee2e84"} Oct 04 11:33:22 crc kubenswrapper[5025]: I1004 11:33:22.140753 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.197583564 podStartE2EDuration="3.140737964s" podCreationTimestamp="2025-10-04 11:33:19 +0000 UTC" firstStartedPulling="2025-10-04 11:33:20.211414681 +0000 UTC m=+3528.636381561" lastFinishedPulling="2025-10-04 11:33:21.154569081 +0000 UTC m=+3529.579535961" observedRunningTime="2025-10-04 11:33:22.132153182 +0000 UTC m=+3530.557120052" watchObservedRunningTime="2025-10-04 11:33:22.140737964 +0000 UTC m=+3530.565704844" Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.393929 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-r5n6d/must-gather-x2tq2"] Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.396887 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/must-gather-x2tq2" Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.398602 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-r5n6d"/"openshift-service-ca.crt" Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.398778 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-r5n6d"/"kube-root-ca.crt" Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.431168 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-r5n6d/must-gather-x2tq2"] Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.581405 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd5dg\" (UniqueName: \"kubernetes.io/projected/2baea1e1-8444-4751-9b61-65fb509d9a91-kube-api-access-pd5dg\") pod \"must-gather-x2tq2\" (UID: \"2baea1e1-8444-4751-9b61-65fb509d9a91\") " pod="openshift-must-gather-r5n6d/must-gather-x2tq2" Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.581502 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2baea1e1-8444-4751-9b61-65fb509d9a91-must-gather-output\") pod \"must-gather-x2tq2\" (UID: \"2baea1e1-8444-4751-9b61-65fb509d9a91\") " pod="openshift-must-gather-r5n6d/must-gather-x2tq2" Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.682834 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd5dg\" (UniqueName: \"kubernetes.io/projected/2baea1e1-8444-4751-9b61-65fb509d9a91-kube-api-access-pd5dg\") pod \"must-gather-x2tq2\" (UID: \"2baea1e1-8444-4751-9b61-65fb509d9a91\") " pod="openshift-must-gather-r5n6d/must-gather-x2tq2" Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.682935 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2baea1e1-8444-4751-9b61-65fb509d9a91-must-gather-output\") pod \"must-gather-x2tq2\" (UID: \"2baea1e1-8444-4751-9b61-65fb509d9a91\") " pod="openshift-must-gather-r5n6d/must-gather-x2tq2" Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.683670 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2baea1e1-8444-4751-9b61-65fb509d9a91-must-gather-output\") pod \"must-gather-x2tq2\" (UID: \"2baea1e1-8444-4751-9b61-65fb509d9a91\") " pod="openshift-must-gather-r5n6d/must-gather-x2tq2" Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.700647 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd5dg\" (UniqueName: \"kubernetes.io/projected/2baea1e1-8444-4751-9b61-65fb509d9a91-kube-api-access-pd5dg\") pod \"must-gather-x2tq2\" (UID: \"2baea1e1-8444-4751-9b61-65fb509d9a91\") " pod="openshift-must-gather-r5n6d/must-gather-x2tq2" Oct 04 11:33:38 crc kubenswrapper[5025]: I1004 11:33:38.718630 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/must-gather-x2tq2" Oct 04 11:33:39 crc kubenswrapper[5025]: I1004 11:33:39.152436 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-r5n6d/must-gather-x2tq2"] Oct 04 11:33:39 crc kubenswrapper[5025]: I1004 11:33:39.301720 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/must-gather-x2tq2" event={"ID":"2baea1e1-8444-4751-9b61-65fb509d9a91","Type":"ContainerStarted","Data":"532334a0c744fcc7c40a7e9d58ac36f22bd6fbaf5262801b0f0e61a2bdcc35d4"} Oct 04 11:33:43 crc kubenswrapper[5025]: I1004 11:33:43.340395 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/must-gather-x2tq2" event={"ID":"2baea1e1-8444-4751-9b61-65fb509d9a91","Type":"ContainerStarted","Data":"5c0f12979f189ee7466257e27700bb07f2789cab9f5d4e8a9d85d10b282fdefb"} Oct 04 11:33:43 crc kubenswrapper[5025]: I1004 11:33:43.341077 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/must-gather-x2tq2" event={"ID":"2baea1e1-8444-4751-9b61-65fb509d9a91","Type":"ContainerStarted","Data":"8e45ed2397e16cae34689d00fca69c20a9b0fb4269c4df6a895ec969f7106b53"} Oct 04 11:33:43 crc kubenswrapper[5025]: I1004 11:33:43.353354 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-r5n6d/must-gather-x2tq2" podStartSLOduration=1.758370378 podStartE2EDuration="5.353332948s" podCreationTimestamp="2025-10-04 11:33:38 +0000 UTC" firstStartedPulling="2025-10-04 11:33:39.165698254 +0000 UTC m=+3547.590665134" lastFinishedPulling="2025-10-04 11:33:42.760660814 +0000 UTC m=+3551.185627704" observedRunningTime="2025-10-04 11:33:43.352619168 +0000 UTC m=+3551.777586048" watchObservedRunningTime="2025-10-04 11:33:43.353332948 +0000 UTC m=+3551.778299828" Oct 04 11:33:44 crc kubenswrapper[5025]: I1004 11:33:44.713200 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:33:44 crc kubenswrapper[5025]: I1004 11:33:44.713649 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:33:46 crc kubenswrapper[5025]: I1004 11:33:46.585325 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-r5n6d/crc-debug-trn9l"] Oct 04 11:33:46 crc kubenswrapper[5025]: I1004 11:33:46.588153 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-trn9l" Oct 04 11:33:46 crc kubenswrapper[5025]: I1004 11:33:46.590149 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-r5n6d"/"default-dockercfg-ghjcz" Oct 04 11:33:46 crc kubenswrapper[5025]: I1004 11:33:46.739715 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x45b\" (UniqueName: \"kubernetes.io/projected/b445aef3-023f-45d9-a6a8-297de0f03524-kube-api-access-7x45b\") pod \"crc-debug-trn9l\" (UID: \"b445aef3-023f-45d9-a6a8-297de0f03524\") " pod="openshift-must-gather-r5n6d/crc-debug-trn9l" Oct 04 11:33:46 crc kubenswrapper[5025]: I1004 11:33:46.739951 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b445aef3-023f-45d9-a6a8-297de0f03524-host\") pod \"crc-debug-trn9l\" (UID: \"b445aef3-023f-45d9-a6a8-297de0f03524\") " pod="openshift-must-gather-r5n6d/crc-debug-trn9l" Oct 04 11:33:46 crc kubenswrapper[5025]: I1004 11:33:46.841907 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x45b\" (UniqueName: \"kubernetes.io/projected/b445aef3-023f-45d9-a6a8-297de0f03524-kube-api-access-7x45b\") pod \"crc-debug-trn9l\" (UID: \"b445aef3-023f-45d9-a6a8-297de0f03524\") " pod="openshift-must-gather-r5n6d/crc-debug-trn9l" Oct 04 11:33:46 crc kubenswrapper[5025]: I1004 11:33:46.842445 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b445aef3-023f-45d9-a6a8-297de0f03524-host\") pod \"crc-debug-trn9l\" (UID: \"b445aef3-023f-45d9-a6a8-297de0f03524\") " pod="openshift-must-gather-r5n6d/crc-debug-trn9l" Oct 04 11:33:46 crc kubenswrapper[5025]: I1004 11:33:46.842582 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b445aef3-023f-45d9-a6a8-297de0f03524-host\") pod \"crc-debug-trn9l\" (UID: \"b445aef3-023f-45d9-a6a8-297de0f03524\") " pod="openshift-must-gather-r5n6d/crc-debug-trn9l" Oct 04 11:33:46 crc kubenswrapper[5025]: I1004 11:33:46.866058 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x45b\" (UniqueName: \"kubernetes.io/projected/b445aef3-023f-45d9-a6a8-297de0f03524-kube-api-access-7x45b\") pod \"crc-debug-trn9l\" (UID: \"b445aef3-023f-45d9-a6a8-297de0f03524\") " pod="openshift-must-gather-r5n6d/crc-debug-trn9l" Oct 04 11:33:46 crc kubenswrapper[5025]: I1004 11:33:46.908413 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-trn9l" Oct 04 11:33:47 crc kubenswrapper[5025]: I1004 11:33:47.379829 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/crc-debug-trn9l" event={"ID":"b445aef3-023f-45d9-a6a8-297de0f03524","Type":"ContainerStarted","Data":"ca5e6cacf003638dc870a2510082f0e9a2a4d0059b90b2ad6cedd3171372a6b9"} Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.336555 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p4q9w"] Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.365636 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4q9w"] Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.365852 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.528439 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97clp\" (UniqueName: \"kubernetes.io/projected/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-kube-api-access-97clp\") pod \"community-operators-p4q9w\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.528627 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-catalog-content\") pod \"community-operators-p4q9w\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.528699 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-utilities\") pod \"community-operators-p4q9w\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.631437 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97clp\" (UniqueName: \"kubernetes.io/projected/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-kube-api-access-97clp\") pod \"community-operators-p4q9w\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.631560 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-catalog-content\") pod \"community-operators-p4q9w\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.631597 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-utilities\") pod \"community-operators-p4q9w\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.632111 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-catalog-content\") pod \"community-operators-p4q9w\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.632149 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-utilities\") pod \"community-operators-p4q9w\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.653535 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97clp\" (UniqueName: \"kubernetes.io/projected/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-kube-api-access-97clp\") pod \"community-operators-p4q9w\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:55 crc kubenswrapper[5025]: I1004 11:33:55.689715 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:33:59 crc kubenswrapper[5025]: I1004 11:33:59.102305 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4q9w"] Oct 04 11:34:00 crc kubenswrapper[5025]: I1004 11:34:00.523986 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q9w" event={"ID":"7363a5d1-17d2-4af3-8327-92bf2b1eadc9","Type":"ContainerStarted","Data":"acb40c7dcc42a82dbccad4ec8ef8aaef08d5799261f0d3a184448f082e20a490"} Oct 04 11:34:02 crc kubenswrapper[5025]: I1004 11:34:02.544663 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/crc-debug-trn9l" event={"ID":"b445aef3-023f-45d9-a6a8-297de0f03524","Type":"ContainerStarted","Data":"3aa189f763a2f320c6265479cf67259fa03a8deec922faa10912071260a2341a"} Oct 04 11:34:02 crc kubenswrapper[5025]: I1004 11:34:02.547371 5025 generic.go:334] "Generic (PLEG): container finished" podID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerID="adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db" exitCode=0 Oct 04 11:34:02 crc kubenswrapper[5025]: I1004 11:34:02.547414 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q9w" event={"ID":"7363a5d1-17d2-4af3-8327-92bf2b1eadc9","Type":"ContainerDied","Data":"adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db"} Oct 04 11:34:02 crc kubenswrapper[5025]: I1004 11:34:02.573061 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-r5n6d/crc-debug-trn9l" podStartSLOduration=3.119040367 podStartE2EDuration="16.573041915s" podCreationTimestamp="2025-10-04 11:33:46 +0000 UTC" firstStartedPulling="2025-10-04 11:33:46.949730168 +0000 UTC m=+3555.374697048" lastFinishedPulling="2025-10-04 11:34:00.403731716 +0000 UTC m=+3568.828698596" observedRunningTime="2025-10-04 11:34:02.558837944 +0000 UTC m=+3570.983804824" watchObservedRunningTime="2025-10-04 11:34:02.573041915 +0000 UTC m=+3570.998008795" Oct 04 11:34:03 crc kubenswrapper[5025]: I1004 11:34:03.556576 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q9w" event={"ID":"7363a5d1-17d2-4af3-8327-92bf2b1eadc9","Type":"ContainerStarted","Data":"36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c"} Oct 04 11:34:04 crc kubenswrapper[5025]: I1004 11:34:04.570444 5025 generic.go:334] "Generic (PLEG): container finished" podID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerID="36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c" exitCode=0 Oct 04 11:34:04 crc kubenswrapper[5025]: I1004 11:34:04.570506 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q9w" event={"ID":"7363a5d1-17d2-4af3-8327-92bf2b1eadc9","Type":"ContainerDied","Data":"36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c"} Oct 04 11:34:05 crc kubenswrapper[5025]: I1004 11:34:05.580294 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q9w" event={"ID":"7363a5d1-17d2-4af3-8327-92bf2b1eadc9","Type":"ContainerStarted","Data":"6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29"} Oct 04 11:34:05 crc kubenswrapper[5025]: I1004 11:34:05.604364 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p4q9w" podStartSLOduration=8.07335508 podStartE2EDuration="10.60434552s" podCreationTimestamp="2025-10-04 11:33:55 +0000 UTC" firstStartedPulling="2025-10-04 11:34:02.550954131 +0000 UTC m=+3570.975921011" lastFinishedPulling="2025-10-04 11:34:05.081944571 +0000 UTC m=+3573.506911451" observedRunningTime="2025-10-04 11:34:05.596222681 +0000 UTC m=+3574.021189561" watchObservedRunningTime="2025-10-04 11:34:05.60434552 +0000 UTC m=+3574.029312400" Oct 04 11:34:05 crc kubenswrapper[5025]: I1004 11:34:05.690871 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:34:05 crc kubenswrapper[5025]: I1004 11:34:05.690941 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:34:06 crc kubenswrapper[5025]: I1004 11:34:06.752035 5025 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-p4q9w" podUID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerName="registry-server" probeResult="failure" output=< Oct 04 11:34:06 crc kubenswrapper[5025]: timeout: failed to connect service ":50051" within 1s Oct 04 11:34:06 crc kubenswrapper[5025]: > Oct 04 11:34:14 crc kubenswrapper[5025]: I1004 11:34:14.713944 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:34:14 crc kubenswrapper[5025]: I1004 11:34:14.714407 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:34:15 crc kubenswrapper[5025]: I1004 11:34:15.738179 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:34:15 crc kubenswrapper[5025]: I1004 11:34:15.791725 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:34:15 crc kubenswrapper[5025]: I1004 11:34:15.973784 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4q9w"] Oct 04 11:34:17 crc kubenswrapper[5025]: I1004 11:34:17.687599 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p4q9w" podUID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerName="registry-server" containerID="cri-o://6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29" gracePeriod=2 Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.199172 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.340729 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-catalog-content\") pod \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.341145 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97clp\" (UniqueName: \"kubernetes.io/projected/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-kube-api-access-97clp\") pod \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.341484 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-utilities\") pod \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\" (UID: \"7363a5d1-17d2-4af3-8327-92bf2b1eadc9\") " Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.342497 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-utilities" (OuterVolumeSpecName: "utilities") pod "7363a5d1-17d2-4af3-8327-92bf2b1eadc9" (UID: "7363a5d1-17d2-4af3-8327-92bf2b1eadc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.357713 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-kube-api-access-97clp" (OuterVolumeSpecName: "kube-api-access-97clp") pod "7363a5d1-17d2-4af3-8327-92bf2b1eadc9" (UID: "7363a5d1-17d2-4af3-8327-92bf2b1eadc9"). InnerVolumeSpecName "kube-api-access-97clp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.415322 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7363a5d1-17d2-4af3-8327-92bf2b1eadc9" (UID: "7363a5d1-17d2-4af3-8327-92bf2b1eadc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.444295 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.444330 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.444341 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97clp\" (UniqueName: \"kubernetes.io/projected/7363a5d1-17d2-4af3-8327-92bf2b1eadc9-kube-api-access-97clp\") on node \"crc\" DevicePath \"\"" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.697816 5025 generic.go:334] "Generic (PLEG): container finished" podID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerID="6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29" exitCode=0 Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.697857 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q9w" event={"ID":"7363a5d1-17d2-4af3-8327-92bf2b1eadc9","Type":"ContainerDied","Data":"6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29"} Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.697863 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4q9w" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.697881 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4q9w" event={"ID":"7363a5d1-17d2-4af3-8327-92bf2b1eadc9","Type":"ContainerDied","Data":"acb40c7dcc42a82dbccad4ec8ef8aaef08d5799261f0d3a184448f082e20a490"} Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.697896 5025 scope.go:117] "RemoveContainer" containerID="6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.722191 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4q9w"] Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.722383 5025 scope.go:117] "RemoveContainer" containerID="36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.732088 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p4q9w"] Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.746910 5025 scope.go:117] "RemoveContainer" containerID="adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.795091 5025 scope.go:117] "RemoveContainer" containerID="6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29" Oct 04 11:34:18 crc kubenswrapper[5025]: E1004 11:34:18.795725 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29\": container with ID starting with 6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29 not found: ID does not exist" containerID="6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.795796 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29"} err="failed to get container status \"6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29\": rpc error: code = NotFound desc = could not find container \"6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29\": container with ID starting with 6f038807b3063a16adcfd98b61792d37e741e36d74a0bd31946554f7a44d6f29 not found: ID does not exist" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.795823 5025 scope.go:117] "RemoveContainer" containerID="36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c" Oct 04 11:34:18 crc kubenswrapper[5025]: E1004 11:34:18.796266 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c\": container with ID starting with 36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c not found: ID does not exist" containerID="36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.796297 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c"} err="failed to get container status \"36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c\": rpc error: code = NotFound desc = could not find container \"36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c\": container with ID starting with 36b1daf0543563e59a1fa092140f78ecd4d6f301a6c553920eb42d6480fa167c not found: ID does not exist" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.796313 5025 scope.go:117] "RemoveContainer" containerID="adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db" Oct 04 11:34:18 crc kubenswrapper[5025]: E1004 11:34:18.796683 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db\": container with ID starting with adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db not found: ID does not exist" containerID="adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db" Oct 04 11:34:18 crc kubenswrapper[5025]: I1004 11:34:18.796708 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db"} err="failed to get container status \"adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db\": rpc error: code = NotFound desc = could not find container \"adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db\": container with ID starting with adf6618401676f4ae105a3781fa0640710bf757b3bffd5fc4d1b5c9392e4c8db not found: ID does not exist" Oct 04 11:34:20 crc kubenswrapper[5025]: I1004 11:34:20.422775 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" path="/var/lib/kubelet/pods/7363a5d1-17d2-4af3-8327-92bf2b1eadc9/volumes" Oct 04 11:34:44 crc kubenswrapper[5025]: I1004 11:34:44.713287 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:34:44 crc kubenswrapper[5025]: I1004 11:34:44.713891 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:34:44 crc kubenswrapper[5025]: I1004 11:34:44.713951 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 11:34:44 crc kubenswrapper[5025]: I1004 11:34:44.714522 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f262faf2a678238fd476697f624cd66122884cd0f3538b657c6e0e36f229e229"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:34:44 crc kubenswrapper[5025]: I1004 11:34:44.714595 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://f262faf2a678238fd476697f624cd66122884cd0f3538b657c6e0e36f229e229" gracePeriod=600 Oct 04 11:34:44 crc kubenswrapper[5025]: I1004 11:34:44.925984 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="f262faf2a678238fd476697f624cd66122884cd0f3538b657c6e0e36f229e229" exitCode=0 Oct 04 11:34:44 crc kubenswrapper[5025]: I1004 11:34:44.926061 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"f262faf2a678238fd476697f624cd66122884cd0f3538b657c6e0e36f229e229"} Oct 04 11:34:44 crc kubenswrapper[5025]: I1004 11:34:44.926126 5025 scope.go:117] "RemoveContainer" containerID="a37e8552e499e334cf69a4361a22a9ae3245f8c5eb59a034f3d22aa97124c807" Oct 04 11:34:45 crc kubenswrapper[5025]: I1004 11:34:45.943173 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51"} Oct 04 11:34:47 crc kubenswrapper[5025]: I1004 11:34:47.769805 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6bfdcdbc7d-swl5f_fd52b707-ec59-4653-974d-9b0963fee564/barbican-api/0.log" Oct 04 11:34:47 crc kubenswrapper[5025]: I1004 11:34:47.813315 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6bfdcdbc7d-swl5f_fd52b707-ec59-4653-974d-9b0963fee564/barbican-api-log/0.log" Oct 04 11:34:48 crc kubenswrapper[5025]: I1004 11:34:48.030713 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-74bc6fc8d4-xzvxt_6acf4ca7-9b65-42cb-8a0d-d09b14292c6e/barbican-keystone-listener/0.log" Oct 04 11:34:48 crc kubenswrapper[5025]: I1004 11:34:48.072924 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-74bc6fc8d4-xzvxt_6acf4ca7-9b65-42cb-8a0d-d09b14292c6e/barbican-keystone-listener-log/0.log" Oct 04 11:34:48 crc kubenswrapper[5025]: I1004 11:34:48.207681 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6db97b4dc9-fxss9_0cfb63d9-275c-4019-8809-7c90883cab5d/barbican-worker/0.log" Oct 04 11:34:48 crc kubenswrapper[5025]: I1004 11:34:48.294940 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6db97b4dc9-fxss9_0cfb63d9-275c-4019-8809-7c90883cab5d/barbican-worker-log/0.log" Oct 04 11:34:48 crc kubenswrapper[5025]: I1004 11:34:48.498401 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7_28e59cf5-a704-4eb4-94e7-84a90ceca38d/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:48 crc kubenswrapper[5025]: I1004 11:34:48.663566 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0baa8e1a-723d-49e2-9f65-0d95550a8633/ceilometer-central-agent/0.log" Oct 04 11:34:48 crc kubenswrapper[5025]: I1004 11:34:48.701907 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0baa8e1a-723d-49e2-9f65-0d95550a8633/ceilometer-notification-agent/0.log" Oct 04 11:34:48 crc kubenswrapper[5025]: I1004 11:34:48.763381 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0baa8e1a-723d-49e2-9f65-0d95550a8633/proxy-httpd/0.log" Oct 04 11:34:48 crc kubenswrapper[5025]: I1004 11:34:48.834254 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0baa8e1a-723d-49e2-9f65-0d95550a8633/sg-core/0.log" Oct 04 11:34:49 crc kubenswrapper[5025]: I1004 11:34:49.027094 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3f3cd68e-1753-4b60-94ba-4c33d2a95034/cinder-api/0.log" Oct 04 11:34:49 crc kubenswrapper[5025]: I1004 11:34:49.041599 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3f3cd68e-1753-4b60-94ba-4c33d2a95034/cinder-api-log/0.log" Oct 04 11:34:49 crc kubenswrapper[5025]: I1004 11:34:49.283697 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6da062c0-b483-4388-8759-84f44afb073b/probe/0.log" Oct 04 11:34:49 crc kubenswrapper[5025]: I1004 11:34:49.288211 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6da062c0-b483-4388-8759-84f44afb073b/cinder-scheduler/0.log" Oct 04 11:34:49 crc kubenswrapper[5025]: I1004 11:34:49.493323 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw_bfac76b8-067d-4dc5-8d95-930cf9f08863/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:49 crc kubenswrapper[5025]: I1004 11:34:49.645276 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl_015bb9a3-af1a-4721-930b-df387aefb2b8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:49 crc kubenswrapper[5025]: I1004 11:34:49.778374 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5_7f1f31a9-c086-48bb-86e3-6db0b697de2e/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:50 crc kubenswrapper[5025]: I1004 11:34:50.007895 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-ts6zc_1529a91c-0604-4357-be99-3305358c188a/init/0.log" Oct 04 11:34:50 crc kubenswrapper[5025]: I1004 11:34:50.255116 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-ts6zc_1529a91c-0604-4357-be99-3305358c188a/init/0.log" Oct 04 11:34:50 crc kubenswrapper[5025]: I1004 11:34:50.290343 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-ts6zc_1529a91c-0604-4357-be99-3305358c188a/dnsmasq-dns/0.log" Oct 04 11:34:50 crc kubenswrapper[5025]: I1004 11:34:50.462819 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v_aca99a65-247d-49eb-8bb8-4016fe9fdfd0/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:50 crc kubenswrapper[5025]: I1004 11:34:50.581236 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_35af90f2-e78a-4e59-a694-72044d020d8f/glance-httpd/0.log" Oct 04 11:34:50 crc kubenswrapper[5025]: I1004 11:34:50.635510 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_35af90f2-e78a-4e59-a694-72044d020d8f/glance-log/0.log" Oct 04 11:34:50 crc kubenswrapper[5025]: I1004 11:34:50.763939 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5393a1c0-fcc5-449d-a4e5-42723c13d83e/glance-httpd/0.log" Oct 04 11:34:50 crc kubenswrapper[5025]: I1004 11:34:50.796750 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5393a1c0-fcc5-449d-a4e5-42723c13d83e/glance-log/0.log" Oct 04 11:34:51 crc kubenswrapper[5025]: I1004 11:34:51.005081 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6559967db4-zfm7d_53944185-08b0-4249-8662-9fdce7cb24cf/horizon/0.log" Oct 04 11:34:51 crc kubenswrapper[5025]: I1004 11:34:51.152797 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8_2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:51 crc kubenswrapper[5025]: I1004 11:34:51.284379 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6559967db4-zfm7d_53944185-08b0-4249-8662-9fdce7cb24cf/horizon-log/0.log" Oct 04 11:34:51 crc kubenswrapper[5025]: I1004 11:34:51.342297 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-kjz6s_5f5a4382-7710-48d0-a476-8464c98612dc/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:51 crc kubenswrapper[5025]: I1004 11:34:51.551108 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29326261-55b5l_fc05eaa2-72fa-4aef-8a86-a1bae9d427c4/keystone-cron/0.log" Oct 04 11:34:51 crc kubenswrapper[5025]: I1004 11:34:51.631838 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6f856fd68f-ckb8t_15a00f7c-91b2-4635-b46b-8ae22b23f14e/keystone-api/0.log" Oct 04 11:34:51 crc kubenswrapper[5025]: I1004 11:34:51.662167 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_16cbe27e-1c01-4140-abbb-351ddb59d907/kube-state-metrics/0.log" Oct 04 11:34:51 crc kubenswrapper[5025]: I1004 11:34:51.821191 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-6k987_67bb3dde-fd47-406d-8862-b365032f6ac9/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:52 crc kubenswrapper[5025]: I1004 11:34:52.190435 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c7b9c48cc-j7mmw_993d2c5b-1d22-4185-a265-2fe9bffc0318/neutron-httpd/0.log" Oct 04 11:34:52 crc kubenswrapper[5025]: I1004 11:34:52.217271 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c7b9c48cc-j7mmw_993d2c5b-1d22-4185-a265-2fe9bffc0318/neutron-api/0.log" Oct 04 11:34:52 crc kubenswrapper[5025]: I1004 11:34:52.396256 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t_b85d77e5-edce-41bf-9b7f-d1cdda6861f3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:53 crc kubenswrapper[5025]: I1004 11:34:53.041275 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9a3f816a-01d3-4823-b6e2-cf24e30a6735/nova-api-log/0.log" Oct 04 11:34:53 crc kubenswrapper[5025]: I1004 11:34:53.131353 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3cfb2082-5742-4e6f-aedf-a6e22e661ad3/nova-cell0-conductor-conductor/0.log" Oct 04 11:34:53 crc kubenswrapper[5025]: I1004 11:34:53.194479 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9a3f816a-01d3-4823-b6e2-cf24e30a6735/nova-api-api/0.log" Oct 04 11:34:53 crc kubenswrapper[5025]: I1004 11:34:53.505723 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ddc5d59e-f1c0-4b4a-ac69-326ffff2038a/nova-cell1-conductor-conductor/0.log" Oct 04 11:34:53 crc kubenswrapper[5025]: I1004 11:34:53.601264 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5e1bccb3-569f-45d0-a778-0fd8b04f8cca/nova-cell1-novncproxy-novncproxy/0.log" Oct 04 11:34:53 crc kubenswrapper[5025]: I1004 11:34:53.802664 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-tqj97_d807eca9-05be-475e-b168-2ab47ab555c1/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:54 crc kubenswrapper[5025]: I1004 11:34:54.068817 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f03a463d-561b-4517-9167-e2c66c8f323f/nova-metadata-log/0.log" Oct 04 11:34:54 crc kubenswrapper[5025]: I1004 11:34:54.444844 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e8754265-dd19-4438-a891-94cf4a6ebe01/nova-scheduler-scheduler/0.log" Oct 04 11:34:54 crc kubenswrapper[5025]: I1004 11:34:54.615146 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_64ee2b80-da98-4572-a2dc-c08ca7933e61/mysql-bootstrap/0.log" Oct 04 11:34:54 crc kubenswrapper[5025]: I1004 11:34:54.861857 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_64ee2b80-da98-4572-a2dc-c08ca7933e61/mysql-bootstrap/0.log" Oct 04 11:34:54 crc kubenswrapper[5025]: I1004 11:34:54.862811 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_64ee2b80-da98-4572-a2dc-c08ca7933e61/galera/0.log" Oct 04 11:34:55 crc kubenswrapper[5025]: I1004 11:34:55.138541 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5df194f4-9f9f-48e1-a4c9-87409b4f2b6e/mysql-bootstrap/0.log" Oct 04 11:34:55 crc kubenswrapper[5025]: I1004 11:34:55.267717 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f03a463d-561b-4517-9167-e2c66c8f323f/nova-metadata-metadata/0.log" Oct 04 11:34:55 crc kubenswrapper[5025]: I1004 11:34:55.350724 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5df194f4-9f9f-48e1-a4c9-87409b4f2b6e/galera/0.log" Oct 04 11:34:55 crc kubenswrapper[5025]: I1004 11:34:55.382048 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5df194f4-9f9f-48e1-a4c9-87409b4f2b6e/mysql-bootstrap/0.log" Oct 04 11:34:55 crc kubenswrapper[5025]: I1004 11:34:55.584536 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_31297488-5835-4cd4-a9c5-7dff9253423a/openstackclient/0.log" Oct 04 11:34:55 crc kubenswrapper[5025]: I1004 11:34:55.738303 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jcfxn_fe21ee2a-4ed1-47aa-90f3-42629a279fd6/ovn-controller/0.log" Oct 04 11:34:55 crc kubenswrapper[5025]: I1004 11:34:55.878374 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-84wmk_2954cd9e-67cf-4083-8ca7-d1e24449314b/openstack-network-exporter/0.log" Oct 04 11:34:56 crc kubenswrapper[5025]: I1004 11:34:56.043053 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c7fwx_d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3/ovsdb-server-init/0.log" Oct 04 11:34:56 crc kubenswrapper[5025]: I1004 11:34:56.278513 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c7fwx_d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3/ovs-vswitchd/0.log" Oct 04 11:34:56 crc kubenswrapper[5025]: I1004 11:34:56.298486 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c7fwx_d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3/ovsdb-server-init/0.log" Oct 04 11:34:56 crc kubenswrapper[5025]: I1004 11:34:56.345921 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c7fwx_d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3/ovsdb-server/0.log" Oct 04 11:34:56 crc kubenswrapper[5025]: I1004 11:34:56.565366 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-42n5d_e3d2c355-fea3-4828-8291-77d0daa90ee5/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:56 crc kubenswrapper[5025]: I1004 11:34:56.715632 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ee98710d-b3c1-4dbf-ab50-ee831c4d622d/openstack-network-exporter/0.log" Oct 04 11:34:56 crc kubenswrapper[5025]: I1004 11:34:56.809666 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ee98710d-b3c1-4dbf-ab50-ee831c4d622d/ovn-northd/0.log" Oct 04 11:34:56 crc kubenswrapper[5025]: I1004 11:34:56.906451 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c1026dd7-48ec-4cb2-aa1d-0e823fefea09/openstack-network-exporter/0.log" Oct 04 11:34:57 crc kubenswrapper[5025]: I1004 11:34:57.033990 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c1026dd7-48ec-4cb2-aa1d-0e823fefea09/ovsdbserver-nb/0.log" Oct 04 11:34:57 crc kubenswrapper[5025]: I1004 11:34:57.152533 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9887ac7f-49c1-4316-92ba-5107f6f9260d/openstack-network-exporter/0.log" Oct 04 11:34:57 crc kubenswrapper[5025]: I1004 11:34:57.225544 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9887ac7f-49c1-4316-92ba-5107f6f9260d/ovsdbserver-sb/0.log" Oct 04 11:34:57 crc kubenswrapper[5025]: I1004 11:34:57.416782 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-67f4db9dc8-qprqb_3a1c727a-b835-42d9-ae4d-81f6ea3018fa/placement-api/0.log" Oct 04 11:34:57 crc kubenswrapper[5025]: I1004 11:34:57.569227 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-67f4db9dc8-qprqb_3a1c727a-b835-42d9-ae4d-81f6ea3018fa/placement-log/0.log" Oct 04 11:34:57 crc kubenswrapper[5025]: I1004 11:34:57.763284 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ec47d2a-5987-4d71-bee2-648b3d664135/setup-container/0.log" Oct 04 11:34:57 crc kubenswrapper[5025]: I1004 11:34:57.977293 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ec47d2a-5987-4d71-bee2-648b3d664135/setup-container/0.log" Oct 04 11:34:58 crc kubenswrapper[5025]: I1004 11:34:58.017823 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ec47d2a-5987-4d71-bee2-648b3d664135/rabbitmq/0.log" Oct 04 11:34:58 crc kubenswrapper[5025]: I1004 11:34:58.204992 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8adc9762-e9bc-427e-87a7-777fb153920c/setup-container/0.log" Oct 04 11:34:58 crc kubenswrapper[5025]: I1004 11:34:58.495535 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8adc9762-e9bc-427e-87a7-777fb153920c/rabbitmq/0.log" Oct 04 11:34:58 crc kubenswrapper[5025]: I1004 11:34:58.511095 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8adc9762-e9bc-427e-87a7-777fb153920c/setup-container/0.log" Oct 04 11:34:58 crc kubenswrapper[5025]: I1004 11:34:58.683827 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5_c9ecdd38-1c7f-4739-b9e5-80ef801564d9/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:58 crc kubenswrapper[5025]: I1004 11:34:58.803046 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-2xpd7_663838d1-caa8-481f-b9fe-b8f51fc6eed9/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:58 crc kubenswrapper[5025]: I1004 11:34:58.932528 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9_cc9937b2-f2f6-4e38-8c69-aaf75422be6a/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:59 crc kubenswrapper[5025]: I1004 11:34:59.133621 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-g8clb_f677e673-14c1-4f54-b7db-15b99f3922d2/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:34:59 crc kubenswrapper[5025]: I1004 11:34:59.313755 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-jv7vr_e4ba226b-2376-4bd2-823c-c8fc00c46369/ssh-known-hosts-edpm-deployment/0.log" Oct 04 11:34:59 crc kubenswrapper[5025]: I1004 11:34:59.535826 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-568f969787-rrbtk_48087914-5acc-46e7-8ce2-382f6fe28d48/proxy-httpd/0.log" Oct 04 11:34:59 crc kubenswrapper[5025]: I1004 11:34:59.571107 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-568f969787-rrbtk_48087914-5acc-46e7-8ce2-382f6fe28d48/proxy-server/0.log" Oct 04 11:34:59 crc kubenswrapper[5025]: I1004 11:34:59.785782 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-wmvlp_49009ba5-e7dd-408f-ad33-a9fa20d156a6/swift-ring-rebalance/0.log" Oct 04 11:34:59 crc kubenswrapper[5025]: I1004 11:34:59.786065 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/account-auditor/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.002246 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/account-replicator/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.044126 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/account-reaper/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.048689 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/account-server/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.196700 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/container-auditor/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.264611 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/container-replicator/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.317426 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/container-server/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.435081 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/container-updater/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.514148 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/object-auditor/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.558904 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/object-expirer/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.682524 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/object-replicator/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.698754 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/object-server/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.793945 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/object-updater/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.844085 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/rsync/0.log" Oct 04 11:35:00 crc kubenswrapper[5025]: I1004 11:35:00.955797 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/swift-recon-cron/0.log" Oct 04 11:35:01 crc kubenswrapper[5025]: I1004 11:35:01.131153 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd_4d9059c6-5fa0-4cc7-a7b5-490f85405f9f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:35:01 crc kubenswrapper[5025]: I1004 11:35:01.288731 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5c783e4c-998b-44ab-956a-de26d5568f90/tempest-tests-tempest-tests-runner/0.log" Oct 04 11:35:01 crc kubenswrapper[5025]: I1004 11:35:01.456573 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_ea4d0d6d-6d8b-45c3-abbb-d69654bc2622/test-operator-logs-container/0.log" Oct 04 11:35:01 crc kubenswrapper[5025]: I1004 11:35:01.632035 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk_9f4db205-f2ac-428f-a424-a238cfa9d0be/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:35:08 crc kubenswrapper[5025]: I1004 11:35:08.932233 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2588531d-38e6-486e-b120-94bbd8a61b4f/memcached/0.log" Oct 04 11:35:54 crc kubenswrapper[5025]: I1004 11:35:54.608321 5025 generic.go:334] "Generic (PLEG): container finished" podID="b445aef3-023f-45d9-a6a8-297de0f03524" containerID="3aa189f763a2f320c6265479cf67259fa03a8deec922faa10912071260a2341a" exitCode=0 Oct 04 11:35:54 crc kubenswrapper[5025]: I1004 11:35:54.608409 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/crc-debug-trn9l" event={"ID":"b445aef3-023f-45d9-a6a8-297de0f03524","Type":"ContainerDied","Data":"3aa189f763a2f320c6265479cf67259fa03a8deec922faa10912071260a2341a"} Oct 04 11:35:55 crc kubenswrapper[5025]: I1004 11:35:55.733980 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-trn9l" Oct 04 11:35:55 crc kubenswrapper[5025]: I1004 11:35:55.768864 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-r5n6d/crc-debug-trn9l"] Oct 04 11:35:55 crc kubenswrapper[5025]: I1004 11:35:55.775951 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-r5n6d/crc-debug-trn9l"] Oct 04 11:35:55 crc kubenswrapper[5025]: I1004 11:35:55.849658 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x45b\" (UniqueName: \"kubernetes.io/projected/b445aef3-023f-45d9-a6a8-297de0f03524-kube-api-access-7x45b\") pod \"b445aef3-023f-45d9-a6a8-297de0f03524\" (UID: \"b445aef3-023f-45d9-a6a8-297de0f03524\") " Oct 04 11:35:55 crc kubenswrapper[5025]: I1004 11:35:55.849898 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b445aef3-023f-45d9-a6a8-297de0f03524-host\") pod \"b445aef3-023f-45d9-a6a8-297de0f03524\" (UID: \"b445aef3-023f-45d9-a6a8-297de0f03524\") " Oct 04 11:35:55 crc kubenswrapper[5025]: I1004 11:35:55.850101 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b445aef3-023f-45d9-a6a8-297de0f03524-host" (OuterVolumeSpecName: "host") pod "b445aef3-023f-45d9-a6a8-297de0f03524" (UID: "b445aef3-023f-45d9-a6a8-297de0f03524"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:35:55 crc kubenswrapper[5025]: I1004 11:35:55.850510 5025 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b445aef3-023f-45d9-a6a8-297de0f03524-host\") on node \"crc\" DevicePath \"\"" Oct 04 11:35:55 crc kubenswrapper[5025]: I1004 11:35:55.857795 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b445aef3-023f-45d9-a6a8-297de0f03524-kube-api-access-7x45b" (OuterVolumeSpecName: "kube-api-access-7x45b") pod "b445aef3-023f-45d9-a6a8-297de0f03524" (UID: "b445aef3-023f-45d9-a6a8-297de0f03524"). InnerVolumeSpecName "kube-api-access-7x45b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:35:55 crc kubenswrapper[5025]: I1004 11:35:55.952268 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x45b\" (UniqueName: \"kubernetes.io/projected/b445aef3-023f-45d9-a6a8-297de0f03524-kube-api-access-7x45b\") on node \"crc\" DevicePath \"\"" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.425200 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b445aef3-023f-45d9-a6a8-297de0f03524" path="/var/lib/kubelet/pods/b445aef3-023f-45d9-a6a8-297de0f03524/volumes" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.629423 5025 scope.go:117] "RemoveContainer" containerID="3aa189f763a2f320c6265479cf67259fa03a8deec922faa10912071260a2341a" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.629480 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-trn9l" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.967493 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-r5n6d/crc-debug-nttzq"] Oct 04 11:35:56 crc kubenswrapper[5025]: E1004 11:35:56.968147 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerName="extract-utilities" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.968169 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerName="extract-utilities" Oct 04 11:35:56 crc kubenswrapper[5025]: E1004 11:35:56.968204 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b445aef3-023f-45d9-a6a8-297de0f03524" containerName="container-00" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.968219 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b445aef3-023f-45d9-a6a8-297de0f03524" containerName="container-00" Oct 04 11:35:56 crc kubenswrapper[5025]: E1004 11:35:56.968238 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerName="extract-content" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.968253 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerName="extract-content" Oct 04 11:35:56 crc kubenswrapper[5025]: E1004 11:35:56.968293 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerName="registry-server" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.968306 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerName="registry-server" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.968686 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="b445aef3-023f-45d9-a6a8-297de0f03524" containerName="container-00" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.968728 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="7363a5d1-17d2-4af3-8327-92bf2b1eadc9" containerName="registry-server" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.969885 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-nttzq" Oct 04 11:35:56 crc kubenswrapper[5025]: I1004 11:35:56.973233 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-r5n6d"/"default-dockercfg-ghjcz" Oct 04 11:35:57 crc kubenswrapper[5025]: I1004 11:35:57.076729 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wptnj\" (UniqueName: \"kubernetes.io/projected/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-kube-api-access-wptnj\") pod \"crc-debug-nttzq\" (UID: \"e4f7bca6-393c-4bc6-9778-2ac285d92bd2\") " pod="openshift-must-gather-r5n6d/crc-debug-nttzq" Oct 04 11:35:57 crc kubenswrapper[5025]: I1004 11:35:57.077178 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-host\") pod \"crc-debug-nttzq\" (UID: \"e4f7bca6-393c-4bc6-9778-2ac285d92bd2\") " pod="openshift-must-gather-r5n6d/crc-debug-nttzq" Oct 04 11:35:57 crc kubenswrapper[5025]: I1004 11:35:57.179277 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wptnj\" (UniqueName: \"kubernetes.io/projected/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-kube-api-access-wptnj\") pod \"crc-debug-nttzq\" (UID: \"e4f7bca6-393c-4bc6-9778-2ac285d92bd2\") " pod="openshift-must-gather-r5n6d/crc-debug-nttzq" Oct 04 11:35:57 crc kubenswrapper[5025]: I1004 11:35:57.179464 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-host\") pod \"crc-debug-nttzq\" (UID: \"e4f7bca6-393c-4bc6-9778-2ac285d92bd2\") " pod="openshift-must-gather-r5n6d/crc-debug-nttzq" Oct 04 11:35:57 crc kubenswrapper[5025]: I1004 11:35:57.179557 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-host\") pod \"crc-debug-nttzq\" (UID: \"e4f7bca6-393c-4bc6-9778-2ac285d92bd2\") " pod="openshift-must-gather-r5n6d/crc-debug-nttzq" Oct 04 11:35:57 crc kubenswrapper[5025]: I1004 11:35:57.214909 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wptnj\" (UniqueName: \"kubernetes.io/projected/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-kube-api-access-wptnj\") pod \"crc-debug-nttzq\" (UID: \"e4f7bca6-393c-4bc6-9778-2ac285d92bd2\") " pod="openshift-must-gather-r5n6d/crc-debug-nttzq" Oct 04 11:35:57 crc kubenswrapper[5025]: I1004 11:35:57.294207 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-nttzq" Oct 04 11:35:57 crc kubenswrapper[5025]: W1004 11:35:57.344037 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4f7bca6_393c_4bc6_9778_2ac285d92bd2.slice/crio-d3dc854ce4939e03f3730ed636cc0b7e0746298935c3deb931dbc0f8ec41ee1a WatchSource:0}: Error finding container d3dc854ce4939e03f3730ed636cc0b7e0746298935c3deb931dbc0f8ec41ee1a: Status 404 returned error can't find the container with id d3dc854ce4939e03f3730ed636cc0b7e0746298935c3deb931dbc0f8ec41ee1a Oct 04 11:35:57 crc kubenswrapper[5025]: I1004 11:35:57.642977 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/crc-debug-nttzq" event={"ID":"e4f7bca6-393c-4bc6-9778-2ac285d92bd2","Type":"ContainerStarted","Data":"f0cc21280f85ac59b6c3950d17b7b815955adce30405f23d34162c183ae336c9"} Oct 04 11:35:57 crc kubenswrapper[5025]: I1004 11:35:57.643317 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/crc-debug-nttzq" event={"ID":"e4f7bca6-393c-4bc6-9778-2ac285d92bd2","Type":"ContainerStarted","Data":"d3dc854ce4939e03f3730ed636cc0b7e0746298935c3deb931dbc0f8ec41ee1a"} Oct 04 11:35:57 crc kubenswrapper[5025]: I1004 11:35:57.672075 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-r5n6d/crc-debug-nttzq" podStartSLOduration=1.672056581 podStartE2EDuration="1.672056581s" podCreationTimestamp="2025-10-04 11:35:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:35:57.657698025 +0000 UTC m=+3686.082664935" watchObservedRunningTime="2025-10-04 11:35:57.672056581 +0000 UTC m=+3686.097023461" Oct 04 11:35:58 crc kubenswrapper[5025]: I1004 11:35:58.657915 5025 generic.go:334] "Generic (PLEG): container finished" podID="e4f7bca6-393c-4bc6-9778-2ac285d92bd2" containerID="f0cc21280f85ac59b6c3950d17b7b815955adce30405f23d34162c183ae336c9" exitCode=0 Oct 04 11:35:58 crc kubenswrapper[5025]: I1004 11:35:58.657999 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/crc-debug-nttzq" event={"ID":"e4f7bca6-393c-4bc6-9778-2ac285d92bd2","Type":"ContainerDied","Data":"f0cc21280f85ac59b6c3950d17b7b815955adce30405f23d34162c183ae336c9"} Oct 04 11:35:59 crc kubenswrapper[5025]: I1004 11:35:59.768852 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-nttzq" Oct 04 11:35:59 crc kubenswrapper[5025]: I1004 11:35:59.820695 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-host\") pod \"e4f7bca6-393c-4bc6-9778-2ac285d92bd2\" (UID: \"e4f7bca6-393c-4bc6-9778-2ac285d92bd2\") " Oct 04 11:35:59 crc kubenswrapper[5025]: I1004 11:35:59.820788 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-host" (OuterVolumeSpecName: "host") pod "e4f7bca6-393c-4bc6-9778-2ac285d92bd2" (UID: "e4f7bca6-393c-4bc6-9778-2ac285d92bd2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:35:59 crc kubenswrapper[5025]: I1004 11:35:59.820926 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wptnj\" (UniqueName: \"kubernetes.io/projected/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-kube-api-access-wptnj\") pod \"e4f7bca6-393c-4bc6-9778-2ac285d92bd2\" (UID: \"e4f7bca6-393c-4bc6-9778-2ac285d92bd2\") " Oct 04 11:35:59 crc kubenswrapper[5025]: I1004 11:35:59.821528 5025 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-host\") on node \"crc\" DevicePath \"\"" Oct 04 11:35:59 crc kubenswrapper[5025]: I1004 11:35:59.831259 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-kube-api-access-wptnj" (OuterVolumeSpecName: "kube-api-access-wptnj") pod "e4f7bca6-393c-4bc6-9778-2ac285d92bd2" (UID: "e4f7bca6-393c-4bc6-9778-2ac285d92bd2"). InnerVolumeSpecName "kube-api-access-wptnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:35:59 crc kubenswrapper[5025]: I1004 11:35:59.922987 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wptnj\" (UniqueName: \"kubernetes.io/projected/e4f7bca6-393c-4bc6-9778-2ac285d92bd2-kube-api-access-wptnj\") on node \"crc\" DevicePath \"\"" Oct 04 11:36:00 crc kubenswrapper[5025]: I1004 11:36:00.676955 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/crc-debug-nttzq" event={"ID":"e4f7bca6-393c-4bc6-9778-2ac285d92bd2","Type":"ContainerDied","Data":"d3dc854ce4939e03f3730ed636cc0b7e0746298935c3deb931dbc0f8ec41ee1a"} Oct 04 11:36:00 crc kubenswrapper[5025]: I1004 11:36:00.676989 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-nttzq" Oct 04 11:36:00 crc kubenswrapper[5025]: I1004 11:36:00.677054 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3dc854ce4939e03f3730ed636cc0b7e0746298935c3deb931dbc0f8ec41ee1a" Oct 04 11:36:04 crc kubenswrapper[5025]: I1004 11:36:04.330601 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-r5n6d/crc-debug-nttzq"] Oct 04 11:36:04 crc kubenswrapper[5025]: I1004 11:36:04.337696 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-r5n6d/crc-debug-nttzq"] Oct 04 11:36:04 crc kubenswrapper[5025]: I1004 11:36:04.426273 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f7bca6-393c-4bc6-9778-2ac285d92bd2" path="/var/lib/kubelet/pods/e4f7bca6-393c-4bc6-9778-2ac285d92bd2/volumes" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.565054 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-r5n6d/crc-debug-9k8h8"] Oct 04 11:36:05 crc kubenswrapper[5025]: E1004 11:36:05.565826 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f7bca6-393c-4bc6-9778-2ac285d92bd2" containerName="container-00" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.565842 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f7bca6-393c-4bc6-9778-2ac285d92bd2" containerName="container-00" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.566218 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f7bca6-393c-4bc6-9778-2ac285d92bd2" containerName="container-00" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.566905 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.571191 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-r5n6d"/"default-dockercfg-ghjcz" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.609458 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf75t\" (UniqueName: \"kubernetes.io/projected/c23a3e1d-90f7-47df-94b9-cceb28fb3148-kube-api-access-lf75t\") pod \"crc-debug-9k8h8\" (UID: \"c23a3e1d-90f7-47df-94b9-cceb28fb3148\") " pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.609606 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c23a3e1d-90f7-47df-94b9-cceb28fb3148-host\") pod \"crc-debug-9k8h8\" (UID: \"c23a3e1d-90f7-47df-94b9-cceb28fb3148\") " pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.711160 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf75t\" (UniqueName: \"kubernetes.io/projected/c23a3e1d-90f7-47df-94b9-cceb28fb3148-kube-api-access-lf75t\") pod \"crc-debug-9k8h8\" (UID: \"c23a3e1d-90f7-47df-94b9-cceb28fb3148\") " pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.711287 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c23a3e1d-90f7-47df-94b9-cceb28fb3148-host\") pod \"crc-debug-9k8h8\" (UID: \"c23a3e1d-90f7-47df-94b9-cceb28fb3148\") " pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.711514 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c23a3e1d-90f7-47df-94b9-cceb28fb3148-host\") pod \"crc-debug-9k8h8\" (UID: \"c23a3e1d-90f7-47df-94b9-cceb28fb3148\") " pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.737508 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf75t\" (UniqueName: \"kubernetes.io/projected/c23a3e1d-90f7-47df-94b9-cceb28fb3148-kube-api-access-lf75t\") pod \"crc-debug-9k8h8\" (UID: \"c23a3e1d-90f7-47df-94b9-cceb28fb3148\") " pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" Oct 04 11:36:05 crc kubenswrapper[5025]: I1004 11:36:05.887609 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" Oct 04 11:36:06 crc kubenswrapper[5025]: I1004 11:36:06.737930 5025 generic.go:334] "Generic (PLEG): container finished" podID="c23a3e1d-90f7-47df-94b9-cceb28fb3148" containerID="f7e5e97758f651dcfe6a7efe6e52b40c43171ff7edf81309fb8e160037cffa60" exitCode=0 Oct 04 11:36:06 crc kubenswrapper[5025]: I1004 11:36:06.738049 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" event={"ID":"c23a3e1d-90f7-47df-94b9-cceb28fb3148","Type":"ContainerDied","Data":"f7e5e97758f651dcfe6a7efe6e52b40c43171ff7edf81309fb8e160037cffa60"} Oct 04 11:36:06 crc kubenswrapper[5025]: I1004 11:36:06.738302 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" event={"ID":"c23a3e1d-90f7-47df-94b9-cceb28fb3148","Type":"ContainerStarted","Data":"1206e96b1196b95d539c448ca696aa538de19473a76f0594bcbeac859b6a820d"} Oct 04 11:36:06 crc kubenswrapper[5025]: I1004 11:36:06.784705 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-r5n6d/crc-debug-9k8h8"] Oct 04 11:36:06 crc kubenswrapper[5025]: I1004 11:36:06.796757 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-r5n6d/crc-debug-9k8h8"] Oct 04 11:36:07 crc kubenswrapper[5025]: I1004 11:36:07.891336 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" Oct 04 11:36:07 crc kubenswrapper[5025]: I1004 11:36:07.953302 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c23a3e1d-90f7-47df-94b9-cceb28fb3148-host\") pod \"c23a3e1d-90f7-47df-94b9-cceb28fb3148\" (UID: \"c23a3e1d-90f7-47df-94b9-cceb28fb3148\") " Oct 04 11:36:07 crc kubenswrapper[5025]: I1004 11:36:07.953386 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf75t\" (UniqueName: \"kubernetes.io/projected/c23a3e1d-90f7-47df-94b9-cceb28fb3148-kube-api-access-lf75t\") pod \"c23a3e1d-90f7-47df-94b9-cceb28fb3148\" (UID: \"c23a3e1d-90f7-47df-94b9-cceb28fb3148\") " Oct 04 11:36:07 crc kubenswrapper[5025]: I1004 11:36:07.953409 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c23a3e1d-90f7-47df-94b9-cceb28fb3148-host" (OuterVolumeSpecName: "host") pod "c23a3e1d-90f7-47df-94b9-cceb28fb3148" (UID: "c23a3e1d-90f7-47df-94b9-cceb28fb3148"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:36:07 crc kubenswrapper[5025]: I1004 11:36:07.953874 5025 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c23a3e1d-90f7-47df-94b9-cceb28fb3148-host\") on node \"crc\" DevicePath \"\"" Oct 04 11:36:07 crc kubenswrapper[5025]: I1004 11:36:07.985084 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c23a3e1d-90f7-47df-94b9-cceb28fb3148-kube-api-access-lf75t" (OuterVolumeSpecName: "kube-api-access-lf75t") pod "c23a3e1d-90f7-47df-94b9-cceb28fb3148" (UID: "c23a3e1d-90f7-47df-94b9-cceb28fb3148"). InnerVolumeSpecName "kube-api-access-lf75t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.055225 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf75t\" (UniqueName: \"kubernetes.io/projected/c23a3e1d-90f7-47df-94b9-cceb28fb3148-kube-api-access-lf75t\") on node \"crc\" DevicePath \"\"" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.403877 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/util/0.log" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.421748 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c23a3e1d-90f7-47df-94b9-cceb28fb3148" path="/var/lib/kubelet/pods/c23a3e1d-90f7-47df-94b9-cceb28fb3148/volumes" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.564673 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/util/0.log" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.592341 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/pull/0.log" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.598812 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/pull/0.log" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.753148 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/util/0.log" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.754624 5025 scope.go:117] "RemoveContainer" containerID="f7e5e97758f651dcfe6a7efe6e52b40c43171ff7edf81309fb8e160037cffa60" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.754672 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/crc-debug-9k8h8" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.763192 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/pull/0.log" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.776045 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/extract/0.log" Oct 04 11:36:08 crc kubenswrapper[5025]: I1004 11:36:08.906527 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-58c4cd55f4-mr64p_91b206eb-240e-41b1-b1eb-df0c649554c9/kube-rbac-proxy/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.013995 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-58c4cd55f4-mr64p_91b206eb-240e-41b1-b1eb-df0c649554c9/manager/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.019891 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-dn4sj_b2f1ef79-fb7a-4d17-ad22-08a2f46df01f/kube-rbac-proxy/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.118583 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-dn4sj_b2f1ef79-fb7a-4d17-ad22-08a2f46df01f/manager/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.181784 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-ppf2w_0ed01e7a-3ef6-4cdd-97c5-ba85a84da593/kube-rbac-proxy/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.229390 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-ppf2w_0ed01e7a-3ef6-4cdd-97c5-ba85a84da593/manager/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.298747 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5dc44df7d5-tx2zp_b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228/kube-rbac-proxy/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.426122 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5dc44df7d5-tx2zp_b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228/manager/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.524813 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-54b4974c45-hqbgk_b2b93be0-ff60-4432-b0db-2ec0e6c605a2/manager/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.533542 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-54b4974c45-hqbgk_b2b93be0-ff60-4432-b0db-2ec0e6c605a2/kube-rbac-proxy/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.627607 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-76d5b87f47-f2blx_6f5b4526-b51d-46e6-900d-492cf48c2710/kube-rbac-proxy/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.726866 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-76d5b87f47-f2blx_6f5b4526-b51d-46e6-900d-492cf48c2710/manager/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.762146 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-mjrsp_36628a66-3fd5-4ca5-a6e4-f6d59009b69f/kube-rbac-proxy/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.925092 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-f9rsb_8f554860-9fb5-4fc0-b795-832b03676469/kube-rbac-proxy/0.log" Oct 04 11:36:09 crc kubenswrapper[5025]: I1004 11:36:09.983557 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-f9rsb_8f554860-9fb5-4fc0-b795-832b03676469/manager/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.022529 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-mjrsp_36628a66-3fd5-4ca5-a6e4-f6d59009b69f/manager/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.120841 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b5ccf6d9c-ms9mz_bd496b6b-d68a-466b-a7de-a11157a5f470/kube-rbac-proxy/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.222115 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b5ccf6d9c-ms9mz_bd496b6b-d68a-466b-a7de-a11157a5f470/manager/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.275548 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-r8csq_3ca07b7a-302b-474c-9aae-584e73350772/manager/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.333827 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-r8csq_3ca07b7a-302b-474c-9aae-584e73350772/kube-rbac-proxy/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.428532 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx_b40c81d1-ba67-411d-b9ed-59a4991ac80f/kube-rbac-proxy/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.468129 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx_b40c81d1-ba67-411d-b9ed-59a4991ac80f/manager/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.594744 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-fbp6d_dc8c4aca-63a0-490e-bf40-9377cdaefefc/kube-rbac-proxy/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.623375 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-fbp6d_dc8c4aca-63a0-490e-bf40-9377cdaefefc/manager/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.707734 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4svdm_b8cb1e86-b4dd-48e9-85c2-6c688d283b7d/kube-rbac-proxy/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.865093 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4svdm_b8cb1e86-b4dd-48e9-85c2-6c688d283b7d/manager/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.876235 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-2llfv_e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc/kube-rbac-proxy/0.log" Oct 04 11:36:10 crc kubenswrapper[5025]: I1004 11:36:10.902759 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-2llfv_e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc/manager/0.log" Oct 04 11:36:11 crc kubenswrapper[5025]: I1004 11:36:11.073366 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx_2e70c89d-b3a2-4b15-90eb-91449857b0a2/manager/0.log" Oct 04 11:36:11 crc kubenswrapper[5025]: I1004 11:36:11.086516 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx_2e70c89d-b3a2-4b15-90eb-91449857b0a2/kube-rbac-proxy/0.log" Oct 04 11:36:11 crc kubenswrapper[5025]: I1004 11:36:11.230032 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8c8588487-9l88t_745ffa84-a65e-48f0-aced-495c5be76951/kube-rbac-proxy/0.log" Oct 04 11:36:11 crc kubenswrapper[5025]: I1004 11:36:11.292953 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-89c6d9c58-mtm9h_a1865149-283a-4933-b0d0-164afd0d1717/kube-rbac-proxy/0.log" Oct 04 11:36:11 crc kubenswrapper[5025]: I1004 11:36:11.555679 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-jjmsm_f3962134-5912-472b-903d-29f945e216e6/registry-server/0.log" Oct 04 11:36:11 crc kubenswrapper[5025]: I1004 11:36:11.629736 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-89c6d9c58-mtm9h_a1865149-283a-4933-b0d0-164afd0d1717/operator/0.log" Oct 04 11:36:11 crc kubenswrapper[5025]: I1004 11:36:11.821677 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6d8b6f9b9-dtnvw_0371bb2b-35e9-4770-9fc4-560d5c2c343f/kube-rbac-proxy/0.log" Oct 04 11:36:11 crc kubenswrapper[5025]: I1004 11:36:11.918530 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6d8b6f9b9-dtnvw_0371bb2b-35e9-4770-9fc4-560d5c2c343f/manager/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.016653 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-f6bvd_81158795-65c7-4a01-b9ba-ad40b7d22582/manager/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.047259 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-f6bvd_81158795-65c7-4a01-b9ba-ad40b7d22582/kube-rbac-proxy/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.166078 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml_a48ece71-fc88-4329-8ad3-fe9db58bb99a/operator/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.312724 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-cl2kh_ad4743c5-3d51-421c-98b8-1463831fb92a/kube-rbac-proxy/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.317049 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-cl2kh_ad4743c5-3d51-421c-98b8-1463831fb92a/manager/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.392490 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8c8588487-9l88t_745ffa84-a65e-48f0-aced-495c5be76951/manager/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.464933 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-jrnhn_85d359e6-3627-4c6b-b974-3599728017a3/kube-rbac-proxy/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.561178 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-jrnhn_85d359e6-3627-4c6b-b974-3599728017a3/manager/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.591045 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-zmt6f_a1cffbcd-c643-4b18-8832-d28767b38fc6/manager/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.638334 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-zmt6f_a1cffbcd-c643-4b18-8832-d28767b38fc6/kube-rbac-proxy/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.762064 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-lfgrl_80fb065c-ce02-4189-8b7a-7bbe78181ee9/manager/0.log" Oct 04 11:36:12 crc kubenswrapper[5025]: I1004 11:36:12.773435 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-lfgrl_80fb065c-ce02-4189-8b7a-7bbe78181ee9/kube-rbac-proxy/0.log" Oct 04 11:36:27 crc kubenswrapper[5025]: I1004 11:36:27.831196 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5k929_4f79bb2f-4697-4f2f-80f5-c95f59b3273f/control-plane-machine-set-operator/0.log" Oct 04 11:36:28 crc kubenswrapper[5025]: I1004 11:36:28.032952 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6mdn4_227aafde-cb49-4cff-b561-d7ba6b84a1cc/kube-rbac-proxy/0.log" Oct 04 11:36:28 crc kubenswrapper[5025]: I1004 11:36:28.041582 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6mdn4_227aafde-cb49-4cff-b561-d7ba6b84a1cc/machine-api-operator/0.log" Oct 04 11:36:40 crc kubenswrapper[5025]: I1004 11:36:40.395258 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-hxgr5_8f18ef59-714b-4044-b76f-44cb4b523a3f/cert-manager-controller/0.log" Oct 04 11:36:40 crc kubenswrapper[5025]: I1004 11:36:40.594201 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-54hps_a413e9a9-6200-4e4a-abb7-939ed023a70f/cert-manager-cainjector/0.log" Oct 04 11:36:40 crc kubenswrapper[5025]: I1004 11:36:40.651813 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-8h2xb_a45d9552-c070-4b0b-99b7-50cfd7acbd28/cert-manager-webhook/0.log" Oct 04 11:36:52 crc kubenswrapper[5025]: I1004 11:36:52.314397 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-mtz64_56180570-2df4-4eee-93dd-d403e7c4a24f/nmstate-console-plugin/0.log" Oct 04 11:36:52 crc kubenswrapper[5025]: I1004 11:36:52.471329 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-ksclx_1c998df0-be41-4019-bba6-5e1b4f58a10e/nmstate-handler/0.log" Oct 04 11:36:52 crc kubenswrapper[5025]: I1004 11:36:52.474936 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-55hjr_7604348c-0bc1-4932-8ff6-00b3098c7326/kube-rbac-proxy/0.log" Oct 04 11:36:52 crc kubenswrapper[5025]: I1004 11:36:52.540574 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-55hjr_7604348c-0bc1-4932-8ff6-00b3098c7326/nmstate-metrics/0.log" Oct 04 11:36:52 crc kubenswrapper[5025]: I1004 11:36:52.655524 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-dvdx6_3b47ddfd-e0dd-4a6f-9879-505058ddc2b3/nmstate-operator/0.log" Oct 04 11:36:52 crc kubenswrapper[5025]: I1004 11:36:52.742591 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-g2ggs_4f852867-9d9b-4255-898c-3b26aabf3243/nmstate-webhook/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.009506 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-jqv9k_24bfff08-7460-44c2-8358-c18faa7933fd/kube-rbac-proxy/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.044059 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-jqv9k_24bfff08-7460-44c2-8358-c18faa7933fd/controller/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.241721 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-frr-files/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.331998 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-reloader/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.333640 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-metrics/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.341746 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-frr-files/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.432957 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-reloader/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.604302 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-frr-files/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.621437 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-reloader/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.626433 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-metrics/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.654275 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-metrics/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.768166 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-frr-files/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.771463 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-reloader/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.806752 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-metrics/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.836750 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/controller/0.log" Oct 04 11:37:08 crc kubenswrapper[5025]: I1004 11:37:08.995547 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/frr-metrics/0.log" Oct 04 11:37:09 crc kubenswrapper[5025]: I1004 11:37:09.004100 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/kube-rbac-proxy/0.log" Oct 04 11:37:09 crc kubenswrapper[5025]: I1004 11:37:09.062315 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/kube-rbac-proxy-frr/0.log" Oct 04 11:37:09 crc kubenswrapper[5025]: I1004 11:37:09.200499 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/reloader/0.log" Oct 04 11:37:09 crc kubenswrapper[5025]: I1004 11:37:09.247405 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-ptxm6_27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79/frr-k8s-webhook-server/0.log" Oct 04 11:37:09 crc kubenswrapper[5025]: I1004 11:37:09.437725 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-66f54ffcff-9lmgs_b87c8f6a-58fe-47e2-822b-ad21c79d0123/manager/0.log" Oct 04 11:37:09 crc kubenswrapper[5025]: I1004 11:37:09.595178 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-764d6c8955-dhmw9_20275cb8-c871-4219-8aee-4f47f917ed72/webhook-server/0.log" Oct 04 11:37:09 crc kubenswrapper[5025]: I1004 11:37:09.820103 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6rf4m_d9c990ef-7b47-4ece-b38d-822182cd7407/kube-rbac-proxy/0.log" Oct 04 11:37:10 crc kubenswrapper[5025]: I1004 11:37:10.212963 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6rf4m_d9c990ef-7b47-4ece-b38d-822182cd7407/speaker/0.log" Oct 04 11:37:10 crc kubenswrapper[5025]: I1004 11:37:10.246771 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/frr/0.log" Oct 04 11:37:14 crc kubenswrapper[5025]: I1004 11:37:14.713336 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:37:14 crc kubenswrapper[5025]: I1004 11:37:14.713987 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:37:21 crc kubenswrapper[5025]: I1004 11:37:21.862385 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/util/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.053184 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/pull/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.062788 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/util/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.086347 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/pull/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.301580 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/util/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.319866 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/pull/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.320492 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/extract/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.497926 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-utilities/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.796106 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-utilities/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.809317 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-content/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.823200 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-content/0.log" Oct 04 11:37:22 crc kubenswrapper[5025]: I1004 11:37:22.979858 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-content/0.log" Oct 04 11:37:23 crc kubenswrapper[5025]: I1004 11:37:23.038159 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-utilities/0.log" Oct 04 11:37:23 crc kubenswrapper[5025]: I1004 11:37:23.197222 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-utilities/0.log" Oct 04 11:37:23 crc kubenswrapper[5025]: I1004 11:37:23.405494 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-utilities/0.log" Oct 04 11:37:23 crc kubenswrapper[5025]: I1004 11:37:23.455411 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/registry-server/0.log" Oct 04 11:37:23 crc kubenswrapper[5025]: I1004 11:37:23.473364 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-content/0.log" Oct 04 11:37:23 crc kubenswrapper[5025]: I1004 11:37:23.484030 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-content/0.log" Oct 04 11:37:23 crc kubenswrapper[5025]: I1004 11:37:23.625938 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-utilities/0.log" Oct 04 11:37:23 crc kubenswrapper[5025]: I1004 11:37:23.708856 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-content/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.116606 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/util/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.191003 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/registry-server/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.257303 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/util/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.299123 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/pull/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.309173 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/pull/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.509552 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/util/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.536889 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/extract/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.540757 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/pull/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.735235 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-d2str_894ba87a-0076-47c2-836d-c972664c8900/marketplace-operator/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.781451 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-utilities/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.926665 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-content/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.926739 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-content/0.log" Oct 04 11:37:24 crc kubenswrapper[5025]: I1004 11:37:24.926873 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-utilities/0.log" Oct 04 11:37:25 crc kubenswrapper[5025]: I1004 11:37:25.110894 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-utilities/0.log" Oct 04 11:37:25 crc kubenswrapper[5025]: I1004 11:37:25.157702 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-content/0.log" Oct 04 11:37:25 crc kubenswrapper[5025]: I1004 11:37:25.288778 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/registry-server/0.log" Oct 04 11:37:25 crc kubenswrapper[5025]: I1004 11:37:25.773411 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-utilities/0.log" Oct 04 11:37:25 crc kubenswrapper[5025]: I1004 11:37:25.994549 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-content/0.log" Oct 04 11:37:26 crc kubenswrapper[5025]: I1004 11:37:26.007672 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-utilities/0.log" Oct 04 11:37:26 crc kubenswrapper[5025]: I1004 11:37:26.058884 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-content/0.log" Oct 04 11:37:26 crc kubenswrapper[5025]: I1004 11:37:26.246359 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-content/0.log" Oct 04 11:37:26 crc kubenswrapper[5025]: I1004 11:37:26.248588 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-utilities/0.log" Oct 04 11:37:26 crc kubenswrapper[5025]: I1004 11:37:26.842643 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/registry-server/0.log" Oct 04 11:37:44 crc kubenswrapper[5025]: I1004 11:37:44.713777 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:37:44 crc kubenswrapper[5025]: I1004 11:37:44.714412 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:37:57 crc kubenswrapper[5025]: E1004 11:37:57.926001 5025 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.162:53722->38.102.83.162:36413: read tcp 38.102.83.162:53722->38.102.83.162:36413: read: connection reset by peer Oct 04 11:38:14 crc kubenswrapper[5025]: I1004 11:38:14.713442 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:38:14 crc kubenswrapper[5025]: I1004 11:38:14.714069 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:38:14 crc kubenswrapper[5025]: I1004 11:38:14.714122 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 11:38:14 crc kubenswrapper[5025]: I1004 11:38:14.714929 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:38:14 crc kubenswrapper[5025]: I1004 11:38:14.714994 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" gracePeriod=600 Oct 04 11:38:14 crc kubenswrapper[5025]: E1004 11:38:14.852051 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:38:14 crc kubenswrapper[5025]: I1004 11:38:14.964381 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" exitCode=0 Oct 04 11:38:14 crc kubenswrapper[5025]: I1004 11:38:14.964441 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51"} Oct 04 11:38:14 crc kubenswrapper[5025]: I1004 11:38:14.964486 5025 scope.go:117] "RemoveContainer" containerID="f262faf2a678238fd476697f624cd66122884cd0f3538b657c6e0e36f229e229" Oct 04 11:38:14 crc kubenswrapper[5025]: I1004 11:38:14.965378 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:38:14 crc kubenswrapper[5025]: E1004 11:38:14.965831 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:38:29 crc kubenswrapper[5025]: I1004 11:38:29.410636 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:38:29 crc kubenswrapper[5025]: E1004 11:38:29.411265 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:38:43 crc kubenswrapper[5025]: I1004 11:38:43.411405 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:38:43 crc kubenswrapper[5025]: E1004 11:38:43.412460 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:38:58 crc kubenswrapper[5025]: I1004 11:38:58.412321 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:38:58 crc kubenswrapper[5025]: E1004 11:38:58.413556 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:39:12 crc kubenswrapper[5025]: I1004 11:39:12.434469 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:39:12 crc kubenswrapper[5025]: E1004 11:39:12.436066 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:39:18 crc kubenswrapper[5025]: I1004 11:39:18.668297 5025 generic.go:334] "Generic (PLEG): container finished" podID="2baea1e1-8444-4751-9b61-65fb509d9a91" containerID="8e45ed2397e16cae34689d00fca69c20a9b0fb4269c4df6a895ec969f7106b53" exitCode=0 Oct 04 11:39:18 crc kubenswrapper[5025]: I1004 11:39:18.668401 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-r5n6d/must-gather-x2tq2" event={"ID":"2baea1e1-8444-4751-9b61-65fb509d9a91","Type":"ContainerDied","Data":"8e45ed2397e16cae34689d00fca69c20a9b0fb4269c4df6a895ec969f7106b53"} Oct 04 11:39:18 crc kubenswrapper[5025]: I1004 11:39:18.669537 5025 scope.go:117] "RemoveContainer" containerID="8e45ed2397e16cae34689d00fca69c20a9b0fb4269c4df6a895ec969f7106b53" Oct 04 11:39:19 crc kubenswrapper[5025]: I1004 11:39:19.433916 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-r5n6d_must-gather-x2tq2_2baea1e1-8444-4751-9b61-65fb509d9a91/gather/0.log" Oct 04 11:39:26 crc kubenswrapper[5025]: I1004 11:39:26.411456 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:39:26 crc kubenswrapper[5025]: E1004 11:39:26.412248 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.186687 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hkzmm"] Oct 04 11:39:28 crc kubenswrapper[5025]: E1004 11:39:28.187638 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c23a3e1d-90f7-47df-94b9-cceb28fb3148" containerName="container-00" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.187660 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="c23a3e1d-90f7-47df-94b9-cceb28fb3148" containerName="container-00" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.187896 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="c23a3e1d-90f7-47df-94b9-cceb28fb3148" containerName="container-00" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.189950 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.200046 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkzmm"] Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.234386 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-r5n6d/must-gather-x2tq2"] Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.234680 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-r5n6d/must-gather-x2tq2" podUID="2baea1e1-8444-4751-9b61-65fb509d9a91" containerName="copy" containerID="cri-o://5c0f12979f189ee7466257e27700bb07f2789cab9f5d4e8a9d85d10b282fdefb" gracePeriod=2 Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.244657 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-r5n6d/must-gather-x2tq2"] Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.385368 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mktmd\" (UniqueName: \"kubernetes.io/projected/abc61412-529b-41ef-90c2-5417b41a7308-kube-api-access-mktmd\") pod \"redhat-marketplace-hkzmm\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.385427 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-utilities\") pod \"redhat-marketplace-hkzmm\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.385492 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-catalog-content\") pod \"redhat-marketplace-hkzmm\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.487230 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-catalog-content\") pod \"redhat-marketplace-hkzmm\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.487689 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-catalog-content\") pod \"redhat-marketplace-hkzmm\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.488111 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mktmd\" (UniqueName: \"kubernetes.io/projected/abc61412-529b-41ef-90c2-5417b41a7308-kube-api-access-mktmd\") pod \"redhat-marketplace-hkzmm\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.488157 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-utilities\") pod \"redhat-marketplace-hkzmm\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.488436 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-utilities\") pod \"redhat-marketplace-hkzmm\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.517828 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mktmd\" (UniqueName: \"kubernetes.io/projected/abc61412-529b-41ef-90c2-5417b41a7308-kube-api-access-mktmd\") pod \"redhat-marketplace-hkzmm\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.520284 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.760206 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-r5n6d_must-gather-x2tq2_2baea1e1-8444-4751-9b61-65fb509d9a91/copy/0.log" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.760744 5025 generic.go:334] "Generic (PLEG): container finished" podID="2baea1e1-8444-4751-9b61-65fb509d9a91" containerID="5c0f12979f189ee7466257e27700bb07f2789cab9f5d4e8a9d85d10b282fdefb" exitCode=143 Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.877937 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-r5n6d_must-gather-x2tq2_2baea1e1-8444-4751-9b61-65fb509d9a91/copy/0.log" Oct 04 11:39:28 crc kubenswrapper[5025]: I1004 11:39:28.880465 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/must-gather-x2tq2" Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:28.999620 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd5dg\" (UniqueName: \"kubernetes.io/projected/2baea1e1-8444-4751-9b61-65fb509d9a91-kube-api-access-pd5dg\") pod \"2baea1e1-8444-4751-9b61-65fb509d9a91\" (UID: \"2baea1e1-8444-4751-9b61-65fb509d9a91\") " Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.000106 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2baea1e1-8444-4751-9b61-65fb509d9a91-must-gather-output\") pod \"2baea1e1-8444-4751-9b61-65fb509d9a91\" (UID: \"2baea1e1-8444-4751-9b61-65fb509d9a91\") " Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.006673 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2baea1e1-8444-4751-9b61-65fb509d9a91-kube-api-access-pd5dg" (OuterVolumeSpecName: "kube-api-access-pd5dg") pod "2baea1e1-8444-4751-9b61-65fb509d9a91" (UID: "2baea1e1-8444-4751-9b61-65fb509d9a91"). InnerVolumeSpecName "kube-api-access-pd5dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.037695 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkzmm"] Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.102141 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd5dg\" (UniqueName: \"kubernetes.io/projected/2baea1e1-8444-4751-9b61-65fb509d9a91-kube-api-access-pd5dg\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.161340 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2baea1e1-8444-4751-9b61-65fb509d9a91-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2baea1e1-8444-4751-9b61-65fb509d9a91" (UID: "2baea1e1-8444-4751-9b61-65fb509d9a91"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.203357 5025 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2baea1e1-8444-4751-9b61-65fb509d9a91-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.771532 5025 generic.go:334] "Generic (PLEG): container finished" podID="abc61412-529b-41ef-90c2-5417b41a7308" containerID="b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae" exitCode=0 Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.771710 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkzmm" event={"ID":"abc61412-529b-41ef-90c2-5417b41a7308","Type":"ContainerDied","Data":"b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae"} Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.771914 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkzmm" event={"ID":"abc61412-529b-41ef-90c2-5417b41a7308","Type":"ContainerStarted","Data":"bebe7fcddb764f5036b9315e69fe6b9a066f41b3603791e4bb14f2969ff6cd8a"} Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.774008 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.778677 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-r5n6d_must-gather-x2tq2_2baea1e1-8444-4751-9b61-65fb509d9a91/copy/0.log" Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.779327 5025 scope.go:117] "RemoveContainer" containerID="5c0f12979f189ee7466257e27700bb07f2789cab9f5d4e8a9d85d10b282fdefb" Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.779464 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-r5n6d/must-gather-x2tq2" Oct 04 11:39:29 crc kubenswrapper[5025]: I1004 11:39:29.807510 5025 scope.go:117] "RemoveContainer" containerID="8e45ed2397e16cae34689d00fca69c20a9b0fb4269c4df6a895ec969f7106b53" Oct 04 11:39:30 crc kubenswrapper[5025]: I1004 11:39:30.430342 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2baea1e1-8444-4751-9b61-65fb509d9a91" path="/var/lib/kubelet/pods/2baea1e1-8444-4751-9b61-65fb509d9a91/volumes" Oct 04 11:39:30 crc kubenswrapper[5025]: I1004 11:39:30.794103 5025 generic.go:334] "Generic (PLEG): container finished" podID="abc61412-529b-41ef-90c2-5417b41a7308" containerID="cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7" exitCode=0 Oct 04 11:39:30 crc kubenswrapper[5025]: I1004 11:39:30.794360 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkzmm" event={"ID":"abc61412-529b-41ef-90c2-5417b41a7308","Type":"ContainerDied","Data":"cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7"} Oct 04 11:39:31 crc kubenswrapper[5025]: I1004 11:39:31.805482 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkzmm" event={"ID":"abc61412-529b-41ef-90c2-5417b41a7308","Type":"ContainerStarted","Data":"0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902"} Oct 04 11:39:31 crc kubenswrapper[5025]: I1004 11:39:31.826302 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hkzmm" podStartSLOduration=2.396338884 podStartE2EDuration="3.826287259s" podCreationTimestamp="2025-10-04 11:39:28 +0000 UTC" firstStartedPulling="2025-10-04 11:39:29.773625495 +0000 UTC m=+3898.198592405" lastFinishedPulling="2025-10-04 11:39:31.2035739 +0000 UTC m=+3899.628540780" observedRunningTime="2025-10-04 11:39:31.824055326 +0000 UTC m=+3900.249022246" watchObservedRunningTime="2025-10-04 11:39:31.826287259 +0000 UTC m=+3900.251254139" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.521285 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.521797 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.553355 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4shns"] Oct 04 11:39:38 crc kubenswrapper[5025]: E1004 11:39:38.553949 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2baea1e1-8444-4751-9b61-65fb509d9a91" containerName="copy" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.553978 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2baea1e1-8444-4751-9b61-65fb509d9a91" containerName="copy" Oct 04 11:39:38 crc kubenswrapper[5025]: E1004 11:39:38.554054 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2baea1e1-8444-4751-9b61-65fb509d9a91" containerName="gather" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.554068 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2baea1e1-8444-4751-9b61-65fb509d9a91" containerName="gather" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.554544 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="2baea1e1-8444-4751-9b61-65fb509d9a91" containerName="gather" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.554582 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="2baea1e1-8444-4751-9b61-65fb509d9a91" containerName="copy" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.569970 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.574293 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4shns"] Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.623356 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.690933 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-utilities\") pod \"redhat-operators-4shns\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.691008 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-catalog-content\") pod \"redhat-operators-4shns\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.691159 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glclp\" (UniqueName: \"kubernetes.io/projected/b0c9a904-4395-44b8-965f-2113f0c37cbf-kube-api-access-glclp\") pod \"redhat-operators-4shns\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.792879 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glclp\" (UniqueName: \"kubernetes.io/projected/b0c9a904-4395-44b8-965f-2113f0c37cbf-kube-api-access-glclp\") pod \"redhat-operators-4shns\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.793108 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-utilities\") pod \"redhat-operators-4shns\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.793187 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-catalog-content\") pod \"redhat-operators-4shns\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.793624 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-utilities\") pod \"redhat-operators-4shns\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.793764 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-catalog-content\") pod \"redhat-operators-4shns\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.825835 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glclp\" (UniqueName: \"kubernetes.io/projected/b0c9a904-4395-44b8-965f-2113f0c37cbf-kube-api-access-glclp\") pod \"redhat-operators-4shns\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.903919 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:38 crc kubenswrapper[5025]: I1004 11:39:38.926554 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:39 crc kubenswrapper[5025]: I1004 11:39:39.418818 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4shns"] Oct 04 11:39:39 crc kubenswrapper[5025]: I1004 11:39:39.884870 5025 generic.go:334] "Generic (PLEG): container finished" podID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerID="1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f" exitCode=0 Oct 04 11:39:39 crc kubenswrapper[5025]: I1004 11:39:39.884967 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4shns" event={"ID":"b0c9a904-4395-44b8-965f-2113f0c37cbf","Type":"ContainerDied","Data":"1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f"} Oct 04 11:39:39 crc kubenswrapper[5025]: I1004 11:39:39.885034 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4shns" event={"ID":"b0c9a904-4395-44b8-965f-2113f0c37cbf","Type":"ContainerStarted","Data":"34fb50e0f130ddee3c80114a07045954dfbb61f947f50afd5fdfb83ae473521b"} Oct 04 11:39:40 crc kubenswrapper[5025]: I1004 11:39:40.942497 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkzmm"] Oct 04 11:39:40 crc kubenswrapper[5025]: I1004 11:39:40.943074 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hkzmm" podUID="abc61412-529b-41ef-90c2-5417b41a7308" containerName="registry-server" containerID="cri-o://0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902" gracePeriod=2 Oct 04 11:39:41 crc kubenswrapper[5025]: E1004 11:39:41.239596 5025 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabc61412_529b_41ef_90c2_5417b41a7308.slice/crio-conmon-0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902.scope\": RecentStats: unable to find data in memory cache]" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.412313 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:39:41 crc kubenswrapper[5025]: E1004 11:39:41.412717 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.442388 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.550331 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mktmd\" (UniqueName: \"kubernetes.io/projected/abc61412-529b-41ef-90c2-5417b41a7308-kube-api-access-mktmd\") pod \"abc61412-529b-41ef-90c2-5417b41a7308\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.550381 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-catalog-content\") pod \"abc61412-529b-41ef-90c2-5417b41a7308\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.550461 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-utilities\") pod \"abc61412-529b-41ef-90c2-5417b41a7308\" (UID: \"abc61412-529b-41ef-90c2-5417b41a7308\") " Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.551543 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-utilities" (OuterVolumeSpecName: "utilities") pod "abc61412-529b-41ef-90c2-5417b41a7308" (UID: "abc61412-529b-41ef-90c2-5417b41a7308"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.553611 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.556711 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abc61412-529b-41ef-90c2-5417b41a7308-kube-api-access-mktmd" (OuterVolumeSpecName: "kube-api-access-mktmd") pod "abc61412-529b-41ef-90c2-5417b41a7308" (UID: "abc61412-529b-41ef-90c2-5417b41a7308"). InnerVolumeSpecName "kube-api-access-mktmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.567114 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abc61412-529b-41ef-90c2-5417b41a7308" (UID: "abc61412-529b-41ef-90c2-5417b41a7308"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.654957 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mktmd\" (UniqueName: \"kubernetes.io/projected/abc61412-529b-41ef-90c2-5417b41a7308-kube-api-access-mktmd\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.655243 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abc61412-529b-41ef-90c2-5417b41a7308-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.905590 5025 generic.go:334] "Generic (PLEG): container finished" podID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerID="1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa" exitCode=0 Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.905700 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4shns" event={"ID":"b0c9a904-4395-44b8-965f-2113f0c37cbf","Type":"ContainerDied","Data":"1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa"} Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.909399 5025 generic.go:334] "Generic (PLEG): container finished" podID="abc61412-529b-41ef-90c2-5417b41a7308" containerID="0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902" exitCode=0 Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.909445 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkzmm" event={"ID":"abc61412-529b-41ef-90c2-5417b41a7308","Type":"ContainerDied","Data":"0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902"} Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.909453 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hkzmm" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.909475 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hkzmm" event={"ID":"abc61412-529b-41ef-90c2-5417b41a7308","Type":"ContainerDied","Data":"bebe7fcddb764f5036b9315e69fe6b9a066f41b3603791e4bb14f2969ff6cd8a"} Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.909498 5025 scope.go:117] "RemoveContainer" containerID="0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.936156 5025 scope.go:117] "RemoveContainer" containerID="cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7" Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.950373 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkzmm"] Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.960074 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hkzmm"] Oct 04 11:39:41 crc kubenswrapper[5025]: I1004 11:39:41.960358 5025 scope.go:117] "RemoveContainer" containerID="b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae" Oct 04 11:39:42 crc kubenswrapper[5025]: I1004 11:39:42.004223 5025 scope.go:117] "RemoveContainer" containerID="0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902" Oct 04 11:39:42 crc kubenswrapper[5025]: E1004 11:39:42.004773 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902\": container with ID starting with 0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902 not found: ID does not exist" containerID="0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902" Oct 04 11:39:42 crc kubenswrapper[5025]: I1004 11:39:42.004802 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902"} err="failed to get container status \"0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902\": rpc error: code = NotFound desc = could not find container \"0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902\": container with ID starting with 0ac2900d41ce96c4d3ea8cb263ffc6c5dde817731af0a303065cbde835d30902 not found: ID does not exist" Oct 04 11:39:42 crc kubenswrapper[5025]: I1004 11:39:42.004823 5025 scope.go:117] "RemoveContainer" containerID="cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7" Oct 04 11:39:42 crc kubenswrapper[5025]: E1004 11:39:42.005261 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7\": container with ID starting with cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7 not found: ID does not exist" containerID="cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7" Oct 04 11:39:42 crc kubenswrapper[5025]: I1004 11:39:42.005317 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7"} err="failed to get container status \"cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7\": rpc error: code = NotFound desc = could not find container \"cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7\": container with ID starting with cde28bb3dfb7dc787f7aa7e53d4d8bd133dacdc7f4573e408594d4c458af4fb7 not found: ID does not exist" Oct 04 11:39:42 crc kubenswrapper[5025]: I1004 11:39:42.005355 5025 scope.go:117] "RemoveContainer" containerID="b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae" Oct 04 11:39:42 crc kubenswrapper[5025]: E1004 11:39:42.005775 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae\": container with ID starting with b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae not found: ID does not exist" containerID="b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae" Oct 04 11:39:42 crc kubenswrapper[5025]: I1004 11:39:42.005805 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae"} err="failed to get container status \"b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae\": rpc error: code = NotFound desc = could not find container \"b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae\": container with ID starting with b4ee5438362f0bf6ebaa5b1b4c9ed9d0690c4873211e66266c0822fffd7658ae not found: ID does not exist" Oct 04 11:39:42 crc kubenswrapper[5025]: I1004 11:39:42.436644 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abc61412-529b-41ef-90c2-5417b41a7308" path="/var/lib/kubelet/pods/abc61412-529b-41ef-90c2-5417b41a7308/volumes" Oct 04 11:39:42 crc kubenswrapper[5025]: I1004 11:39:42.919709 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4shns" event={"ID":"b0c9a904-4395-44b8-965f-2113f0c37cbf","Type":"ContainerStarted","Data":"8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b"} Oct 04 11:39:42 crc kubenswrapper[5025]: I1004 11:39:42.945108 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4shns" podStartSLOduration=2.44723554 podStartE2EDuration="4.945077474s" podCreationTimestamp="2025-10-04 11:39:38 +0000 UTC" firstStartedPulling="2025-10-04 11:39:39.887118362 +0000 UTC m=+3908.312085242" lastFinishedPulling="2025-10-04 11:39:42.384960266 +0000 UTC m=+3910.809927176" observedRunningTime="2025-10-04 11:39:42.934676278 +0000 UTC m=+3911.359643178" watchObservedRunningTime="2025-10-04 11:39:42.945077474 +0000 UTC m=+3911.370044394" Oct 04 11:39:48 crc kubenswrapper[5025]: I1004 11:39:48.904786 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:48 crc kubenswrapper[5025]: I1004 11:39:48.905480 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:48 crc kubenswrapper[5025]: I1004 11:39:48.985322 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:49 crc kubenswrapper[5025]: I1004 11:39:49.057267 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:49 crc kubenswrapper[5025]: I1004 11:39:49.222300 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4shns"] Oct 04 11:39:51 crc kubenswrapper[5025]: I1004 11:39:51.018636 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4shns" podUID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerName="registry-server" containerID="cri-o://8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b" gracePeriod=2 Oct 04 11:39:51 crc kubenswrapper[5025]: I1004 11:39:51.525430 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:51 crc kubenswrapper[5025]: I1004 11:39:51.662960 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-utilities\") pod \"b0c9a904-4395-44b8-965f-2113f0c37cbf\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " Oct 04 11:39:51 crc kubenswrapper[5025]: I1004 11:39:51.663300 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-catalog-content\") pod \"b0c9a904-4395-44b8-965f-2113f0c37cbf\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " Oct 04 11:39:51 crc kubenswrapper[5025]: I1004 11:39:51.663418 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glclp\" (UniqueName: \"kubernetes.io/projected/b0c9a904-4395-44b8-965f-2113f0c37cbf-kube-api-access-glclp\") pod \"b0c9a904-4395-44b8-965f-2113f0c37cbf\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " Oct 04 11:39:51 crc kubenswrapper[5025]: I1004 11:39:51.663732 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-utilities" (OuterVolumeSpecName: "utilities") pod "b0c9a904-4395-44b8-965f-2113f0c37cbf" (UID: "b0c9a904-4395-44b8-965f-2113f0c37cbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:39:51 crc kubenswrapper[5025]: I1004 11:39:51.663939 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:51 crc kubenswrapper[5025]: I1004 11:39:51.688050 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0c9a904-4395-44b8-965f-2113f0c37cbf-kube-api-access-glclp" (OuterVolumeSpecName: "kube-api-access-glclp") pod "b0c9a904-4395-44b8-965f-2113f0c37cbf" (UID: "b0c9a904-4395-44b8-965f-2113f0c37cbf"). InnerVolumeSpecName "kube-api-access-glclp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:39:51 crc kubenswrapper[5025]: I1004 11:39:51.766423 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glclp\" (UniqueName: \"kubernetes.io/projected/b0c9a904-4395-44b8-965f-2113f0c37cbf-kube-api-access-glclp\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.038368 5025 generic.go:334] "Generic (PLEG): container finished" podID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerID="8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b" exitCode=0 Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.038445 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4shns" event={"ID":"b0c9a904-4395-44b8-965f-2113f0c37cbf","Type":"ContainerDied","Data":"8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b"} Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.038497 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4shns" event={"ID":"b0c9a904-4395-44b8-965f-2113f0c37cbf","Type":"ContainerDied","Data":"34fb50e0f130ddee3c80114a07045954dfbb61f947f50afd5fdfb83ae473521b"} Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.038528 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4shns" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.038537 5025 scope.go:117] "RemoveContainer" containerID="8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.067657 5025 scope.go:117] "RemoveContainer" containerID="1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.096233 5025 scope.go:117] "RemoveContainer" containerID="1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.151504 5025 scope.go:117] "RemoveContainer" containerID="8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b" Oct 04 11:39:52 crc kubenswrapper[5025]: E1004 11:39:52.151905 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b\": container with ID starting with 8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b not found: ID does not exist" containerID="8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.151943 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b"} err="failed to get container status \"8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b\": rpc error: code = NotFound desc = could not find container \"8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b\": container with ID starting with 8fa238d21e04cc6b4862329f6f4e55df3aa8c83e39591002d221f04f127a8b9b not found: ID does not exist" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.151977 5025 scope.go:117] "RemoveContainer" containerID="1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa" Oct 04 11:39:52 crc kubenswrapper[5025]: E1004 11:39:52.152287 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa\": container with ID starting with 1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa not found: ID does not exist" containerID="1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.152321 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa"} err="failed to get container status \"1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa\": rpc error: code = NotFound desc = could not find container \"1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa\": container with ID starting with 1548a93938e4d70608ac81740712d1f1b3c6ba7f578fe7028a921c6cd024fefa not found: ID does not exist" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.152341 5025 scope.go:117] "RemoveContainer" containerID="1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f" Oct 04 11:39:52 crc kubenswrapper[5025]: E1004 11:39:52.152817 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f\": container with ID starting with 1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f not found: ID does not exist" containerID="1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.152863 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f"} err="failed to get container status \"1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f\": rpc error: code = NotFound desc = could not find container \"1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f\": container with ID starting with 1e6dc736eef8222546572a71f298e46e38f073d2bf2e6b57e4cc60ddb4324f6f not found: ID does not exist" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.664153 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0c9a904-4395-44b8-965f-2113f0c37cbf" (UID: "b0c9a904-4395-44b8-965f-2113f0c37cbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.665212 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-catalog-content\") pod \"b0c9a904-4395-44b8-965f-2113f0c37cbf\" (UID: \"b0c9a904-4395-44b8-965f-2113f0c37cbf\") " Oct 04 11:39:52 crc kubenswrapper[5025]: W1004 11:39:52.665390 5025 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b0c9a904-4395-44b8-965f-2113f0c37cbf/volumes/kubernetes.io~empty-dir/catalog-content Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.665434 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0c9a904-4395-44b8-965f-2113f0c37cbf" (UID: "b0c9a904-4395-44b8-965f-2113f0c37cbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.665748 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0c9a904-4395-44b8-965f-2113f0c37cbf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.973937 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4shns"] Oct 04 11:39:52 crc kubenswrapper[5025]: I1004 11:39:52.981906 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4shns"] Oct 04 11:39:54 crc kubenswrapper[5025]: I1004 11:39:54.413522 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:39:54 crc kubenswrapper[5025]: E1004 11:39:54.416205 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:39:54 crc kubenswrapper[5025]: I1004 11:39:54.443433 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0c9a904-4395-44b8-965f-2113f0c37cbf" path="/var/lib/kubelet/pods/b0c9a904-4395-44b8-965f-2113f0c37cbf/volumes" Oct 04 11:40:05 crc kubenswrapper[5025]: I1004 11:40:05.411981 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:40:05 crc kubenswrapper[5025]: E1004 11:40:05.412807 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:40:16 crc kubenswrapper[5025]: I1004 11:40:16.411385 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:40:16 crc kubenswrapper[5025]: E1004 11:40:16.412165 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.549478 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5vxlr/must-gather-r79tc"] Oct 04 11:40:24 crc kubenswrapper[5025]: E1004 11:40:24.551140 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc61412-529b-41ef-90c2-5417b41a7308" containerName="extract-utilities" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.551229 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc61412-529b-41ef-90c2-5417b41a7308" containerName="extract-utilities" Oct 04 11:40:24 crc kubenswrapper[5025]: E1004 11:40:24.551294 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc61412-529b-41ef-90c2-5417b41a7308" containerName="registry-server" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.551373 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc61412-529b-41ef-90c2-5417b41a7308" containerName="registry-server" Oct 04 11:40:24 crc kubenswrapper[5025]: E1004 11:40:24.551435 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc61412-529b-41ef-90c2-5417b41a7308" containerName="extract-content" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.551491 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc61412-529b-41ef-90c2-5417b41a7308" containerName="extract-content" Oct 04 11:40:24 crc kubenswrapper[5025]: E1004 11:40:24.551651 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerName="extract-content" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.551711 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerName="extract-content" Oct 04 11:40:24 crc kubenswrapper[5025]: E1004 11:40:24.551776 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerName="extract-utilities" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.551833 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerName="extract-utilities" Oct 04 11:40:24 crc kubenswrapper[5025]: E1004 11:40:24.551898 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerName="registry-server" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.551956 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerName="registry-server" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.552231 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="abc61412-529b-41ef-90c2-5417b41a7308" containerName="registry-server" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.552329 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0c9a904-4395-44b8-965f-2113f0c37cbf" containerName="registry-server" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.553424 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/must-gather-r79tc" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.557294 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-5vxlr"/"default-dockercfg-wqnkd" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.557387 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5vxlr"/"openshift-service-ca.crt" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.557544 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-5vxlr"/"kube-root-ca.crt" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.587495 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5vxlr/must-gather-r79tc"] Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.612150 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-must-gather-output\") pod \"must-gather-r79tc\" (UID: \"c51b4c7e-c4de-4022-9965-0cc6c67bc31e\") " pod="openshift-must-gather-5vxlr/must-gather-r79tc" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.612315 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzqjj\" (UniqueName: \"kubernetes.io/projected/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-kube-api-access-dzqjj\") pod \"must-gather-r79tc\" (UID: \"c51b4c7e-c4de-4022-9965-0cc6c67bc31e\") " pod="openshift-must-gather-5vxlr/must-gather-r79tc" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.713533 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzqjj\" (UniqueName: \"kubernetes.io/projected/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-kube-api-access-dzqjj\") pod \"must-gather-r79tc\" (UID: \"c51b4c7e-c4de-4022-9965-0cc6c67bc31e\") " pod="openshift-must-gather-5vxlr/must-gather-r79tc" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.713601 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-must-gather-output\") pod \"must-gather-r79tc\" (UID: \"c51b4c7e-c4de-4022-9965-0cc6c67bc31e\") " pod="openshift-must-gather-5vxlr/must-gather-r79tc" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.714055 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-must-gather-output\") pod \"must-gather-r79tc\" (UID: \"c51b4c7e-c4de-4022-9965-0cc6c67bc31e\") " pod="openshift-must-gather-5vxlr/must-gather-r79tc" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.735304 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzqjj\" (UniqueName: \"kubernetes.io/projected/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-kube-api-access-dzqjj\") pod \"must-gather-r79tc\" (UID: \"c51b4c7e-c4de-4022-9965-0cc6c67bc31e\") " pod="openshift-must-gather-5vxlr/must-gather-r79tc" Oct 04 11:40:24 crc kubenswrapper[5025]: I1004 11:40:24.869922 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/must-gather-r79tc" Oct 04 11:40:25 crc kubenswrapper[5025]: W1004 11:40:25.365709 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc51b4c7e_c4de_4022_9965_0cc6c67bc31e.slice/crio-ffc968388e1105be6cd54f8afd4e94ca03d044ddce1c9b32ba863f65128c2ad5 WatchSource:0}: Error finding container ffc968388e1105be6cd54f8afd4e94ca03d044ddce1c9b32ba863f65128c2ad5: Status 404 returned error can't find the container with id ffc968388e1105be6cd54f8afd4e94ca03d044ddce1c9b32ba863f65128c2ad5 Oct 04 11:40:25 crc kubenswrapper[5025]: I1004 11:40:25.367998 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-5vxlr/must-gather-r79tc"] Oct 04 11:40:25 crc kubenswrapper[5025]: I1004 11:40:25.377438 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/must-gather-r79tc" event={"ID":"c51b4c7e-c4de-4022-9965-0cc6c67bc31e","Type":"ContainerStarted","Data":"ffc968388e1105be6cd54f8afd4e94ca03d044ddce1c9b32ba863f65128c2ad5"} Oct 04 11:40:26 crc kubenswrapper[5025]: I1004 11:40:26.386497 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/must-gather-r79tc" event={"ID":"c51b4c7e-c4de-4022-9965-0cc6c67bc31e","Type":"ContainerStarted","Data":"7ee702bbe93c8df4806c249c16e94cf87ad779b7ee37be0028d6057fbe4f4d56"} Oct 04 11:40:26 crc kubenswrapper[5025]: I1004 11:40:26.386748 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/must-gather-r79tc" event={"ID":"c51b4c7e-c4de-4022-9965-0cc6c67bc31e","Type":"ContainerStarted","Data":"9c120ab46ead71029270acaf51a8b5336e6feaaed740c204ac8ab3256e113c44"} Oct 04 11:40:28 crc kubenswrapper[5025]: E1004 11:40:28.338738 5025 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.162:56886->38.102.83.162:36413: write tcp 38.102.83.162:56886->38.102.83.162:36413: write: broken pipe Oct 04 11:40:28 crc kubenswrapper[5025]: I1004 11:40:28.980603 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5vxlr/must-gather-r79tc" podStartSLOduration=4.980582839 podStartE2EDuration="4.980582839s" podCreationTimestamp="2025-10-04 11:40:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:40:26.406003853 +0000 UTC m=+3954.830970753" watchObservedRunningTime="2025-10-04 11:40:28.980582839 +0000 UTC m=+3957.405549729" Oct 04 11:40:28 crc kubenswrapper[5025]: I1004 11:40:28.986317 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5vxlr/crc-debug-2fhgn"] Oct 04 11:40:28 crc kubenswrapper[5025]: I1004 11:40:28.988459 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" Oct 04 11:40:29 crc kubenswrapper[5025]: I1004 11:40:29.105143 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/175819d4-659b-4e74-836a-491645fc7363-host\") pod \"crc-debug-2fhgn\" (UID: \"175819d4-659b-4e74-836a-491645fc7363\") " pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" Oct 04 11:40:29 crc kubenswrapper[5025]: I1004 11:40:29.105278 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46brs\" (UniqueName: \"kubernetes.io/projected/175819d4-659b-4e74-836a-491645fc7363-kube-api-access-46brs\") pod \"crc-debug-2fhgn\" (UID: \"175819d4-659b-4e74-836a-491645fc7363\") " pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" Oct 04 11:40:29 crc kubenswrapper[5025]: I1004 11:40:29.206605 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46brs\" (UniqueName: \"kubernetes.io/projected/175819d4-659b-4e74-836a-491645fc7363-kube-api-access-46brs\") pod \"crc-debug-2fhgn\" (UID: \"175819d4-659b-4e74-836a-491645fc7363\") " pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" Oct 04 11:40:29 crc kubenswrapper[5025]: I1004 11:40:29.206772 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/175819d4-659b-4e74-836a-491645fc7363-host\") pod \"crc-debug-2fhgn\" (UID: \"175819d4-659b-4e74-836a-491645fc7363\") " pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" Oct 04 11:40:29 crc kubenswrapper[5025]: I1004 11:40:29.206921 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/175819d4-659b-4e74-836a-491645fc7363-host\") pod \"crc-debug-2fhgn\" (UID: \"175819d4-659b-4e74-836a-491645fc7363\") " pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" Oct 04 11:40:29 crc kubenswrapper[5025]: I1004 11:40:29.228802 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46brs\" (UniqueName: \"kubernetes.io/projected/175819d4-659b-4e74-836a-491645fc7363-kube-api-access-46brs\") pod \"crc-debug-2fhgn\" (UID: \"175819d4-659b-4e74-836a-491645fc7363\") " pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" Oct 04 11:40:29 crc kubenswrapper[5025]: I1004 11:40:29.308228 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" Oct 04 11:40:29 crc kubenswrapper[5025]: W1004 11:40:29.345932 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod175819d4_659b_4e74_836a_491645fc7363.slice/crio-42487941ba4b5993107836589a9eae34d7b1b2bcfdea5309afde1a0f9135ffbe WatchSource:0}: Error finding container 42487941ba4b5993107836589a9eae34d7b1b2bcfdea5309afde1a0f9135ffbe: Status 404 returned error can't find the container with id 42487941ba4b5993107836589a9eae34d7b1b2bcfdea5309afde1a0f9135ffbe Oct 04 11:40:29 crc kubenswrapper[5025]: I1004 11:40:29.415000 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" event={"ID":"175819d4-659b-4e74-836a-491645fc7363","Type":"ContainerStarted","Data":"42487941ba4b5993107836589a9eae34d7b1b2bcfdea5309afde1a0f9135ffbe"} Oct 04 11:40:30 crc kubenswrapper[5025]: I1004 11:40:30.424351 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" event={"ID":"175819d4-659b-4e74-836a-491645fc7363","Type":"ContainerStarted","Data":"d56e864224e31d091d0d1d07fd68b9db2c60024d7b9e883e39745d927b71cc4b"} Oct 04 11:40:31 crc kubenswrapper[5025]: I1004 11:40:31.411736 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:40:31 crc kubenswrapper[5025]: E1004 11:40:31.412514 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:40:44 crc kubenswrapper[5025]: I1004 11:40:44.411640 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:40:44 crc kubenswrapper[5025]: E1004 11:40:44.412494 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:40:55 crc kubenswrapper[5025]: I1004 11:40:55.411184 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:40:55 crc kubenswrapper[5025]: E1004 11:40:55.412130 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:41:08 crc kubenswrapper[5025]: I1004 11:41:08.411642 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:41:08 crc kubenswrapper[5025]: E1004 11:41:08.412448 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:41:21 crc kubenswrapper[5025]: I1004 11:41:21.412372 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:41:21 crc kubenswrapper[5025]: E1004 11:41:21.412994 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:41:30 crc kubenswrapper[5025]: I1004 11:41:30.561619 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6bfdcdbc7d-swl5f_fd52b707-ec59-4653-974d-9b0963fee564/barbican-api-log/0.log" Oct 04 11:41:30 crc kubenswrapper[5025]: I1004 11:41:30.573730 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-6bfdcdbc7d-swl5f_fd52b707-ec59-4653-974d-9b0963fee564/barbican-api/0.log" Oct 04 11:41:30 crc kubenswrapper[5025]: I1004 11:41:30.756173 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-74bc6fc8d4-xzvxt_6acf4ca7-9b65-42cb-8a0d-d09b14292c6e/barbican-keystone-listener/0.log" Oct 04 11:41:30 crc kubenswrapper[5025]: I1004 11:41:30.841173 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-74bc6fc8d4-xzvxt_6acf4ca7-9b65-42cb-8a0d-d09b14292c6e/barbican-keystone-listener-log/0.log" Oct 04 11:41:30 crc kubenswrapper[5025]: I1004 11:41:30.998329 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6db97b4dc9-fxss9_0cfb63d9-275c-4019-8809-7c90883cab5d/barbican-worker/0.log" Oct 04 11:41:31 crc kubenswrapper[5025]: I1004 11:41:31.061904 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6db97b4dc9-fxss9_0cfb63d9-275c-4019-8809-7c90883cab5d/barbican-worker-log/0.log" Oct 04 11:41:31 crc kubenswrapper[5025]: I1004 11:41:31.204125 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-mjzd7_28e59cf5-a704-4eb4-94e7-84a90ceca38d/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:31 crc kubenswrapper[5025]: I1004 11:41:31.930960 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0baa8e1a-723d-49e2-9f65-0d95550a8633/ceilometer-central-agent/0.log" Oct 04 11:41:31 crc kubenswrapper[5025]: I1004 11:41:31.976045 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0baa8e1a-723d-49e2-9f65-0d95550a8633/ceilometer-notification-agent/0.log" Oct 04 11:41:32 crc kubenswrapper[5025]: I1004 11:41:32.085989 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0baa8e1a-723d-49e2-9f65-0d95550a8633/proxy-httpd/0.log" Oct 04 11:41:32 crc kubenswrapper[5025]: I1004 11:41:32.100230 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0baa8e1a-723d-49e2-9f65-0d95550a8633/sg-core/0.log" Oct 04 11:41:32 crc kubenswrapper[5025]: I1004 11:41:32.362646 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3f3cd68e-1753-4b60-94ba-4c33d2a95034/cinder-api/0.log" Oct 04 11:41:32 crc kubenswrapper[5025]: I1004 11:41:32.408866 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3f3cd68e-1753-4b60-94ba-4c33d2a95034/cinder-api-log/0.log" Oct 04 11:41:32 crc kubenswrapper[5025]: I1004 11:41:32.592372 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6da062c0-b483-4388-8759-84f44afb073b/cinder-scheduler/0.log" Oct 04 11:41:32 crc kubenswrapper[5025]: I1004 11:41:32.684729 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6da062c0-b483-4388-8759-84f44afb073b/probe/0.log" Oct 04 11:41:32 crc kubenswrapper[5025]: I1004 11:41:32.829426 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-bdnxw_bfac76b8-067d-4dc5-8d95-930cf9f08863/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:33 crc kubenswrapper[5025]: I1004 11:41:33.556499 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-jxpzl_015bb9a3-af1a-4721-930b-df387aefb2b8/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:33 crc kubenswrapper[5025]: I1004 11:41:33.561083 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-lhvs5_7f1f31a9-c086-48bb-86e3-6db0b697de2e/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:33 crc kubenswrapper[5025]: I1004 11:41:33.771700 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-ts6zc_1529a91c-0604-4357-be99-3305358c188a/init/0.log" Oct 04 11:41:33 crc kubenswrapper[5025]: I1004 11:41:33.951164 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-ts6zc_1529a91c-0604-4357-be99-3305358c188a/init/0.log" Oct 04 11:41:33 crc kubenswrapper[5025]: I1004 11:41:33.975373 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-ts6zc_1529a91c-0604-4357-be99-3305358c188a/dnsmasq-dns/0.log" Oct 04 11:41:34 crc kubenswrapper[5025]: I1004 11:41:34.058953 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-w4m9v_aca99a65-247d-49eb-8bb8-4016fe9fdfd0/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:34 crc kubenswrapper[5025]: I1004 11:41:34.213147 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_35af90f2-e78a-4e59-a694-72044d020d8f/glance-log/0.log" Oct 04 11:41:34 crc kubenswrapper[5025]: I1004 11:41:34.243509 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_35af90f2-e78a-4e59-a694-72044d020d8f/glance-httpd/0.log" Oct 04 11:41:34 crc kubenswrapper[5025]: I1004 11:41:34.381035 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5393a1c0-fcc5-449d-a4e5-42723c13d83e/glance-log/0.log" Oct 04 11:41:34 crc kubenswrapper[5025]: I1004 11:41:34.411061 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:41:34 crc kubenswrapper[5025]: E1004 11:41:34.411381 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:41:34 crc kubenswrapper[5025]: I1004 11:41:34.417434 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_5393a1c0-fcc5-449d-a4e5-42723c13d83e/glance-httpd/0.log" Oct 04 11:41:34 crc kubenswrapper[5025]: I1004 11:41:34.540340 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6559967db4-zfm7d_53944185-08b0-4249-8662-9fdce7cb24cf/horizon/0.log" Oct 04 11:41:34 crc kubenswrapper[5025]: I1004 11:41:34.801608 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-pc8k8_2a823d9e-3ca1-4c78-b9b4-58281e3f1aa9/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:35 crc kubenswrapper[5025]: I1004 11:41:35.027969 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6559967db4-zfm7d_53944185-08b0-4249-8662-9fdce7cb24cf/horizon-log/0.log" Oct 04 11:41:35 crc kubenswrapper[5025]: I1004 11:41:35.067810 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-kjz6s_5f5a4382-7710-48d0-a476-8464c98612dc/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:35 crc kubenswrapper[5025]: I1004 11:41:35.274192 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29326261-55b5l_fc05eaa2-72fa-4aef-8a86-a1bae9d427c4/keystone-cron/0.log" Oct 04 11:41:35 crc kubenswrapper[5025]: I1004 11:41:35.275237 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6f856fd68f-ckb8t_15a00f7c-91b2-4635-b46b-8ae22b23f14e/keystone-api/0.log" Oct 04 11:41:35 crc kubenswrapper[5025]: I1004 11:41:35.407170 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_16cbe27e-1c01-4140-abbb-351ddb59d907/kube-state-metrics/0.log" Oct 04 11:41:35 crc kubenswrapper[5025]: I1004 11:41:35.482567 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-6k987_67bb3dde-fd47-406d-8862-b365032f6ac9/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:35 crc kubenswrapper[5025]: I1004 11:41:35.780475 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c7b9c48cc-j7mmw_993d2c5b-1d22-4185-a265-2fe9bffc0318/neutron-httpd/0.log" Oct 04 11:41:35 crc kubenswrapper[5025]: I1004 11:41:35.783411 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c7b9c48cc-j7mmw_993d2c5b-1d22-4185-a265-2fe9bffc0318/neutron-api/0.log" Oct 04 11:41:36 crc kubenswrapper[5025]: I1004 11:41:36.008457 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-kb57t_b85d77e5-edce-41bf-9b7f-d1cdda6861f3/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:36 crc kubenswrapper[5025]: I1004 11:41:36.513448 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9a3f816a-01d3-4823-b6e2-cf24e30a6735/nova-api-log/0.log" Oct 04 11:41:36 crc kubenswrapper[5025]: I1004 11:41:36.683456 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3cfb2082-5742-4e6f-aedf-a6e22e661ad3/nova-cell0-conductor-conductor/0.log" Oct 04 11:41:36 crc kubenswrapper[5025]: I1004 11:41:36.876046 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_9a3f816a-01d3-4823-b6e2-cf24e30a6735/nova-api-api/0.log" Oct 04 11:41:37 crc kubenswrapper[5025]: I1004 11:41:37.075942 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ddc5d59e-f1c0-4b4a-ac69-326ffff2038a/nova-cell1-conductor-conductor/0.log" Oct 04 11:41:37 crc kubenswrapper[5025]: I1004 11:41:37.210878 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_5e1bccb3-569f-45d0-a778-0fd8b04f8cca/nova-cell1-novncproxy-novncproxy/0.log" Oct 04 11:41:37 crc kubenswrapper[5025]: I1004 11:41:37.318940 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-tqj97_d807eca9-05be-475e-b168-2ab47ab555c1/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:37 crc kubenswrapper[5025]: I1004 11:41:37.485590 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f03a463d-561b-4517-9167-e2c66c8f323f/nova-metadata-log/0.log" Oct 04 11:41:37 crc kubenswrapper[5025]: I1004 11:41:37.952322 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_e8754265-dd19-4438-a891-94cf4a6ebe01/nova-scheduler-scheduler/0.log" Oct 04 11:41:38 crc kubenswrapper[5025]: I1004 11:41:38.049482 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_64ee2b80-da98-4572-a2dc-c08ca7933e61/mysql-bootstrap/0.log" Oct 04 11:41:38 crc kubenswrapper[5025]: I1004 11:41:38.266355 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_64ee2b80-da98-4572-a2dc-c08ca7933e61/mysql-bootstrap/0.log" Oct 04 11:41:38 crc kubenswrapper[5025]: I1004 11:41:38.274564 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_64ee2b80-da98-4572-a2dc-c08ca7933e61/galera/0.log" Oct 04 11:41:38 crc kubenswrapper[5025]: I1004 11:41:38.467299 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5df194f4-9f9f-48e1-a4c9-87409b4f2b6e/mysql-bootstrap/0.log" Oct 04 11:41:38 crc kubenswrapper[5025]: I1004 11:41:38.680864 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5df194f4-9f9f-48e1-a4c9-87409b4f2b6e/mysql-bootstrap/0.log" Oct 04 11:41:38 crc kubenswrapper[5025]: I1004 11:41:38.724027 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_5df194f4-9f9f-48e1-a4c9-87409b4f2b6e/galera/0.log" Oct 04 11:41:38 crc kubenswrapper[5025]: I1004 11:41:38.955708 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_f03a463d-561b-4517-9167-e2c66c8f323f/nova-metadata-metadata/0.log" Oct 04 11:41:38 crc kubenswrapper[5025]: I1004 11:41:38.957287 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_31297488-5835-4cd4-a9c5-7dff9253423a/openstackclient/0.log" Oct 04 11:41:39 crc kubenswrapper[5025]: I1004 11:41:39.166932 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jcfxn_fe21ee2a-4ed1-47aa-90f3-42629a279fd6/ovn-controller/0.log" Oct 04 11:41:39 crc kubenswrapper[5025]: I1004 11:41:39.346596 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-84wmk_2954cd9e-67cf-4083-8ca7-d1e24449314b/openstack-network-exporter/0.log" Oct 04 11:41:39 crc kubenswrapper[5025]: I1004 11:41:39.449541 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c7fwx_d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3/ovsdb-server-init/0.log" Oct 04 11:41:39 crc kubenswrapper[5025]: I1004 11:41:39.646469 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c7fwx_d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3/ovsdb-server-init/0.log" Oct 04 11:41:39 crc kubenswrapper[5025]: I1004 11:41:39.708709 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c7fwx_d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3/ovsdb-server/0.log" Oct 04 11:41:39 crc kubenswrapper[5025]: I1004 11:41:39.713509 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c7fwx_d7ac829f-ad5d-4ec0-9cb9-c500f657e0b3/ovs-vswitchd/0.log" Oct 04 11:41:39 crc kubenswrapper[5025]: I1004 11:41:39.935235 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-42n5d_e3d2c355-fea3-4828-8291-77d0daa90ee5/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:40 crc kubenswrapper[5025]: I1004 11:41:40.069302 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ee98710d-b3c1-4dbf-ab50-ee831c4d622d/openstack-network-exporter/0.log" Oct 04 11:41:40 crc kubenswrapper[5025]: I1004 11:41:40.087113 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ee98710d-b3c1-4dbf-ab50-ee831c4d622d/ovn-northd/0.log" Oct 04 11:41:40 crc kubenswrapper[5025]: I1004 11:41:40.271752 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c1026dd7-48ec-4cb2-aa1d-0e823fefea09/openstack-network-exporter/0.log" Oct 04 11:41:40 crc kubenswrapper[5025]: I1004 11:41:40.336357 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_c1026dd7-48ec-4cb2-aa1d-0e823fefea09/ovsdbserver-nb/0.log" Oct 04 11:41:40 crc kubenswrapper[5025]: I1004 11:41:40.444925 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9887ac7f-49c1-4316-92ba-5107f6f9260d/openstack-network-exporter/0.log" Oct 04 11:41:40 crc kubenswrapper[5025]: I1004 11:41:40.515270 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_9887ac7f-49c1-4316-92ba-5107f6f9260d/ovsdbserver-sb/0.log" Oct 04 11:41:40 crc kubenswrapper[5025]: I1004 11:41:40.757589 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-67f4db9dc8-qprqb_3a1c727a-b835-42d9-ae4d-81f6ea3018fa/placement-api/0.log" Oct 04 11:41:40 crc kubenswrapper[5025]: I1004 11:41:40.797603 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-67f4db9dc8-qprqb_3a1c727a-b835-42d9-ae4d-81f6ea3018fa/placement-log/0.log" Oct 04 11:41:40 crc kubenswrapper[5025]: I1004 11:41:40.999789 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ec47d2a-5987-4d71-bee2-648b3d664135/setup-container/0.log" Oct 04 11:41:41 crc kubenswrapper[5025]: I1004 11:41:41.596941 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ec47d2a-5987-4d71-bee2-648b3d664135/rabbitmq/0.log" Oct 04 11:41:41 crc kubenswrapper[5025]: I1004 11:41:41.635755 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ec47d2a-5987-4d71-bee2-648b3d664135/setup-container/0.log" Oct 04 11:41:41 crc kubenswrapper[5025]: I1004 11:41:41.810770 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8adc9762-e9bc-427e-87a7-777fb153920c/setup-container/0.log" Oct 04 11:41:41 crc kubenswrapper[5025]: I1004 11:41:41.999384 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8adc9762-e9bc-427e-87a7-777fb153920c/setup-container/0.log" Oct 04 11:41:42 crc kubenswrapper[5025]: I1004 11:41:42.009267 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8adc9762-e9bc-427e-87a7-777fb153920c/rabbitmq/0.log" Oct 04 11:41:42 crc kubenswrapper[5025]: I1004 11:41:42.226290 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-cfrn5_c9ecdd38-1c7f-4739-b9e5-80ef801564d9/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:42 crc kubenswrapper[5025]: I1004 11:41:42.366266 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-2xpd7_663838d1-caa8-481f-b9fe-b8f51fc6eed9/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:42 crc kubenswrapper[5025]: I1004 11:41:42.522214 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-5gwt9_cc9937b2-f2f6-4e38-8c69-aaf75422be6a/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:43 crc kubenswrapper[5025]: I1004 11:41:43.216258 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-jv7vr_e4ba226b-2376-4bd2-823c-c8fc00c46369/ssh-known-hosts-edpm-deployment/0.log" Oct 04 11:41:43 crc kubenswrapper[5025]: I1004 11:41:43.268846 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-g8clb_f677e673-14c1-4f54-b7db-15b99f3922d2/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:43 crc kubenswrapper[5025]: I1004 11:41:43.549563 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-568f969787-rrbtk_48087914-5acc-46e7-8ce2-382f6fe28d48/proxy-server/0.log" Oct 04 11:41:43 crc kubenswrapper[5025]: I1004 11:41:43.629781 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-568f969787-rrbtk_48087914-5acc-46e7-8ce2-382f6fe28d48/proxy-httpd/0.log" Oct 04 11:41:43 crc kubenswrapper[5025]: I1004 11:41:43.746152 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-wmvlp_49009ba5-e7dd-408f-ad33-a9fa20d156a6/swift-ring-rebalance/0.log" Oct 04 11:41:43 crc kubenswrapper[5025]: I1004 11:41:43.917897 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/account-auditor/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.003917 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/account-reaper/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.172177 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/account-server/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.185413 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/container-auditor/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.198067 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/account-replicator/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.363371 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/container-server/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.371901 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/container-updater/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.377342 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/container-replicator/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.541202 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/object-expirer/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.593437 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/object-auditor/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.638873 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/object-replicator/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.799006 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/object-server/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.849091 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/object-updater/0.log" Oct 04 11:41:44 crc kubenswrapper[5025]: I1004 11:41:44.946356 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/rsync/0.log" Oct 04 11:41:45 crc kubenswrapper[5025]: I1004 11:41:45.009006 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_369e99d7-ddff-4bfb-9bea-a382aa69eb3e/swift-recon-cron/0.log" Oct 04 11:41:45 crc kubenswrapper[5025]: I1004 11:41:45.134926 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-bbcmd_4d9059c6-5fa0-4cc7-a7b5-490f85405f9f/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:45 crc kubenswrapper[5025]: I1004 11:41:45.280506 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5c783e4c-998b-44ab-956a-de26d5568f90/tempest-tests-tempest-tests-runner/0.log" Oct 04 11:41:45 crc kubenswrapper[5025]: I1004 11:41:45.425641 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_ea4d0d6d-6d8b-45c3-abbb-d69654bc2622/test-operator-logs-container/0.log" Oct 04 11:41:45 crc kubenswrapper[5025]: I1004 11:41:45.577344 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-2d2vk_9f4db205-f2ac-428f-a424-a238cfa9d0be/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 04 11:41:46 crc kubenswrapper[5025]: I1004 11:41:46.410945 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:41:46 crc kubenswrapper[5025]: E1004 11:41:46.411202 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:41:54 crc kubenswrapper[5025]: I1004 11:41:54.402134 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_2588531d-38e6-486e-b120-94bbd8a61b4f/memcached/0.log" Oct 04 11:42:00 crc kubenswrapper[5025]: I1004 11:42:00.411183 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:42:00 crc kubenswrapper[5025]: E1004 11:42:00.411822 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.029879 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" podStartSLOduration=102.02986283 podStartE2EDuration="1m42.02986283s" podCreationTimestamp="2025-10-04 11:40:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:40:30.43817631 +0000 UTC m=+3958.863143220" watchObservedRunningTime="2025-10-04 11:42:10.02986283 +0000 UTC m=+4058.454829710" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.034905 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9qvd5"] Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.037124 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.068932 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9qvd5"] Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.117561 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl5g5\" (UniqueName: \"kubernetes.io/projected/2bf77702-e805-4b2a-a8db-c8aa803c4e28-kube-api-access-wl5g5\") pod \"certified-operators-9qvd5\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.117952 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-utilities\") pod \"certified-operators-9qvd5\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.118424 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-catalog-content\") pod \"certified-operators-9qvd5\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.220752 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl5g5\" (UniqueName: \"kubernetes.io/projected/2bf77702-e805-4b2a-a8db-c8aa803c4e28-kube-api-access-wl5g5\") pod \"certified-operators-9qvd5\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.220830 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-utilities\") pod \"certified-operators-9qvd5\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.220893 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-catalog-content\") pod \"certified-operators-9qvd5\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.221630 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-utilities\") pod \"certified-operators-9qvd5\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.221705 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-catalog-content\") pod \"certified-operators-9qvd5\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.245709 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl5g5\" (UniqueName: \"kubernetes.io/projected/2bf77702-e805-4b2a-a8db-c8aa803c4e28-kube-api-access-wl5g5\") pod \"certified-operators-9qvd5\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.372035 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:10 crc kubenswrapper[5025]: I1004 11:42:10.857948 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9qvd5"] Oct 04 11:42:11 crc kubenswrapper[5025]: I1004 11:42:11.321683 5025 generic.go:334] "Generic (PLEG): container finished" podID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerID="97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c" exitCode=0 Oct 04 11:42:11 crc kubenswrapper[5025]: I1004 11:42:11.321760 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9qvd5" event={"ID":"2bf77702-e805-4b2a-a8db-c8aa803c4e28","Type":"ContainerDied","Data":"97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c"} Oct 04 11:42:11 crc kubenswrapper[5025]: I1004 11:42:11.322037 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9qvd5" event={"ID":"2bf77702-e805-4b2a-a8db-c8aa803c4e28","Type":"ContainerStarted","Data":"642e54c2d3955861b790b3da7e528252aff874641bf9ac4e4944e68defb17893"} Oct 04 11:42:12 crc kubenswrapper[5025]: I1004 11:42:12.435114 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:42:12 crc kubenswrapper[5025]: E1004 11:42:12.436446 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:42:13 crc kubenswrapper[5025]: I1004 11:42:13.351146 5025 generic.go:334] "Generic (PLEG): container finished" podID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerID="2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198" exitCode=0 Oct 04 11:42:13 crc kubenswrapper[5025]: I1004 11:42:13.351194 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9qvd5" event={"ID":"2bf77702-e805-4b2a-a8db-c8aa803c4e28","Type":"ContainerDied","Data":"2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198"} Oct 04 11:42:14 crc kubenswrapper[5025]: I1004 11:42:14.363185 5025 generic.go:334] "Generic (PLEG): container finished" podID="175819d4-659b-4e74-836a-491645fc7363" containerID="d56e864224e31d091d0d1d07fd68b9db2c60024d7b9e883e39745d927b71cc4b" exitCode=0 Oct 04 11:42:14 crc kubenswrapper[5025]: I1004 11:42:14.363240 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" event={"ID":"175819d4-659b-4e74-836a-491645fc7363","Type":"ContainerDied","Data":"d56e864224e31d091d0d1d07fd68b9db2c60024d7b9e883e39745d927b71cc4b"} Oct 04 11:42:14 crc kubenswrapper[5025]: I1004 11:42:14.366203 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9qvd5" event={"ID":"2bf77702-e805-4b2a-a8db-c8aa803c4e28","Type":"ContainerStarted","Data":"8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7"} Oct 04 11:42:14 crc kubenswrapper[5025]: I1004 11:42:14.401815 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9qvd5" podStartSLOduration=1.870417622 podStartE2EDuration="4.401793149s" podCreationTimestamp="2025-10-04 11:42:10 +0000 UTC" firstStartedPulling="2025-10-04 11:42:11.32425211 +0000 UTC m=+4059.749218990" lastFinishedPulling="2025-10-04 11:42:13.855627587 +0000 UTC m=+4062.280594517" observedRunningTime="2025-10-04 11:42:14.399159894 +0000 UTC m=+4062.824126804" watchObservedRunningTime="2025-10-04 11:42:14.401793149 +0000 UTC m=+4062.826760039" Oct 04 11:42:15 crc kubenswrapper[5025]: I1004 11:42:15.471125 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" Oct 04 11:42:15 crc kubenswrapper[5025]: I1004 11:42:15.505174 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5vxlr/crc-debug-2fhgn"] Oct 04 11:42:15 crc kubenswrapper[5025]: I1004 11:42:15.513732 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5vxlr/crc-debug-2fhgn"] Oct 04 11:42:15 crc kubenswrapper[5025]: I1004 11:42:15.622664 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/175819d4-659b-4e74-836a-491645fc7363-host\") pod \"175819d4-659b-4e74-836a-491645fc7363\" (UID: \"175819d4-659b-4e74-836a-491645fc7363\") " Oct 04 11:42:15 crc kubenswrapper[5025]: I1004 11:42:15.622828 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46brs\" (UniqueName: \"kubernetes.io/projected/175819d4-659b-4e74-836a-491645fc7363-kube-api-access-46brs\") pod \"175819d4-659b-4e74-836a-491645fc7363\" (UID: \"175819d4-659b-4e74-836a-491645fc7363\") " Oct 04 11:42:15 crc kubenswrapper[5025]: I1004 11:42:15.622943 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/175819d4-659b-4e74-836a-491645fc7363-host" (OuterVolumeSpecName: "host") pod "175819d4-659b-4e74-836a-491645fc7363" (UID: "175819d4-659b-4e74-836a-491645fc7363"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:42:15 crc kubenswrapper[5025]: I1004 11:42:15.624095 5025 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/175819d4-659b-4e74-836a-491645fc7363-host\") on node \"crc\" DevicePath \"\"" Oct 04 11:42:15 crc kubenswrapper[5025]: I1004 11:42:15.630953 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/175819d4-659b-4e74-836a-491645fc7363-kube-api-access-46brs" (OuterVolumeSpecName: "kube-api-access-46brs") pod "175819d4-659b-4e74-836a-491645fc7363" (UID: "175819d4-659b-4e74-836a-491645fc7363"). InnerVolumeSpecName "kube-api-access-46brs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:42:15 crc kubenswrapper[5025]: I1004 11:42:15.725980 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46brs\" (UniqueName: \"kubernetes.io/projected/175819d4-659b-4e74-836a-491645fc7363-kube-api-access-46brs\") on node \"crc\" DevicePath \"\"" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.383830 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42487941ba4b5993107836589a9eae34d7b1b2bcfdea5309afde1a0f9135ffbe" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.383899 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-2fhgn" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.421208 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="175819d4-659b-4e74-836a-491645fc7363" path="/var/lib/kubelet/pods/175819d4-659b-4e74-836a-491645fc7363/volumes" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.685949 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5vxlr/crc-debug-2rlhg"] Oct 04 11:42:16 crc kubenswrapper[5025]: E1004 11:42:16.687381 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175819d4-659b-4e74-836a-491645fc7363" containerName="container-00" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.687463 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="175819d4-659b-4e74-836a-491645fc7363" containerName="container-00" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.687692 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="175819d4-659b-4e74-836a-491645fc7363" containerName="container-00" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.688477 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.746396 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/47502e79-30b3-4af6-91fe-22658be4eb32-host\") pod \"crc-debug-2rlhg\" (UID: \"47502e79-30b3-4af6-91fe-22658be4eb32\") " pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.746662 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c62k\" (UniqueName: \"kubernetes.io/projected/47502e79-30b3-4af6-91fe-22658be4eb32-kube-api-access-8c62k\") pod \"crc-debug-2rlhg\" (UID: \"47502e79-30b3-4af6-91fe-22658be4eb32\") " pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.848773 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/47502e79-30b3-4af6-91fe-22658be4eb32-host\") pod \"crc-debug-2rlhg\" (UID: \"47502e79-30b3-4af6-91fe-22658be4eb32\") " pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.848862 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c62k\" (UniqueName: \"kubernetes.io/projected/47502e79-30b3-4af6-91fe-22658be4eb32-kube-api-access-8c62k\") pod \"crc-debug-2rlhg\" (UID: \"47502e79-30b3-4af6-91fe-22658be4eb32\") " pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.848915 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/47502e79-30b3-4af6-91fe-22658be4eb32-host\") pod \"crc-debug-2rlhg\" (UID: \"47502e79-30b3-4af6-91fe-22658be4eb32\") " pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" Oct 04 11:42:16 crc kubenswrapper[5025]: I1004 11:42:16.882256 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c62k\" (UniqueName: \"kubernetes.io/projected/47502e79-30b3-4af6-91fe-22658be4eb32-kube-api-access-8c62k\") pod \"crc-debug-2rlhg\" (UID: \"47502e79-30b3-4af6-91fe-22658be4eb32\") " pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" Oct 04 11:42:17 crc kubenswrapper[5025]: I1004 11:42:17.015209 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" Oct 04 11:42:17 crc kubenswrapper[5025]: W1004 11:42:17.054070 5025 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47502e79_30b3_4af6_91fe_22658be4eb32.slice/crio-39ce06265bcdf3687f8b67739d07d7ac6a016ae16d6d679060c04613558adc5c WatchSource:0}: Error finding container 39ce06265bcdf3687f8b67739d07d7ac6a016ae16d6d679060c04613558adc5c: Status 404 returned error can't find the container with id 39ce06265bcdf3687f8b67739d07d7ac6a016ae16d6d679060c04613558adc5c Oct 04 11:42:17 crc kubenswrapper[5025]: I1004 11:42:17.395423 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" event={"ID":"47502e79-30b3-4af6-91fe-22658be4eb32","Type":"ContainerStarted","Data":"1ab75886086e08a7525864238226e0a122343991d47c127e09864875591811c1"} Oct 04 11:42:17 crc kubenswrapper[5025]: I1004 11:42:17.395960 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" event={"ID":"47502e79-30b3-4af6-91fe-22658be4eb32","Type":"ContainerStarted","Data":"39ce06265bcdf3687f8b67739d07d7ac6a016ae16d6d679060c04613558adc5c"} Oct 04 11:42:17 crc kubenswrapper[5025]: I1004 11:42:17.408635 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" podStartSLOduration=1.408623776 podStartE2EDuration="1.408623776s" podCreationTimestamp="2025-10-04 11:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-04 11:42:17.406365292 +0000 UTC m=+4065.831332172" watchObservedRunningTime="2025-10-04 11:42:17.408623776 +0000 UTC m=+4065.833590656" Oct 04 11:42:18 crc kubenswrapper[5025]: I1004 11:42:18.409770 5025 generic.go:334] "Generic (PLEG): container finished" podID="47502e79-30b3-4af6-91fe-22658be4eb32" containerID="1ab75886086e08a7525864238226e0a122343991d47c127e09864875591811c1" exitCode=0 Oct 04 11:42:18 crc kubenswrapper[5025]: I1004 11:42:18.409808 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" event={"ID":"47502e79-30b3-4af6-91fe-22658be4eb32","Type":"ContainerDied","Data":"1ab75886086e08a7525864238226e0a122343991d47c127e09864875591811c1"} Oct 04 11:42:19 crc kubenswrapper[5025]: I1004 11:42:19.531371 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" Oct 04 11:42:19 crc kubenswrapper[5025]: I1004 11:42:19.593330 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/47502e79-30b3-4af6-91fe-22658be4eb32-host\") pod \"47502e79-30b3-4af6-91fe-22658be4eb32\" (UID: \"47502e79-30b3-4af6-91fe-22658be4eb32\") " Oct 04 11:42:19 crc kubenswrapper[5025]: I1004 11:42:19.593394 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/47502e79-30b3-4af6-91fe-22658be4eb32-host" (OuterVolumeSpecName: "host") pod "47502e79-30b3-4af6-91fe-22658be4eb32" (UID: "47502e79-30b3-4af6-91fe-22658be4eb32"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:42:19 crc kubenswrapper[5025]: I1004 11:42:19.593488 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c62k\" (UniqueName: \"kubernetes.io/projected/47502e79-30b3-4af6-91fe-22658be4eb32-kube-api-access-8c62k\") pod \"47502e79-30b3-4af6-91fe-22658be4eb32\" (UID: \"47502e79-30b3-4af6-91fe-22658be4eb32\") " Oct 04 11:42:19 crc kubenswrapper[5025]: I1004 11:42:19.594170 5025 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/47502e79-30b3-4af6-91fe-22658be4eb32-host\") on node \"crc\" DevicePath \"\"" Oct 04 11:42:19 crc kubenswrapper[5025]: I1004 11:42:19.600287 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47502e79-30b3-4af6-91fe-22658be4eb32-kube-api-access-8c62k" (OuterVolumeSpecName: "kube-api-access-8c62k") pod "47502e79-30b3-4af6-91fe-22658be4eb32" (UID: "47502e79-30b3-4af6-91fe-22658be4eb32"). InnerVolumeSpecName "kube-api-access-8c62k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:42:19 crc kubenswrapper[5025]: I1004 11:42:19.696044 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c62k\" (UniqueName: \"kubernetes.io/projected/47502e79-30b3-4af6-91fe-22658be4eb32-kube-api-access-8c62k\") on node \"crc\" DevicePath \"\"" Oct 04 11:42:20 crc kubenswrapper[5025]: I1004 11:42:20.372962 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:20 crc kubenswrapper[5025]: I1004 11:42:20.373031 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:20 crc kubenswrapper[5025]: I1004 11:42:20.426297 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:20 crc kubenswrapper[5025]: I1004 11:42:20.431861 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" Oct 04 11:42:20 crc kubenswrapper[5025]: I1004 11:42:20.432160 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/crc-debug-2rlhg" event={"ID":"47502e79-30b3-4af6-91fe-22658be4eb32","Type":"ContainerDied","Data":"39ce06265bcdf3687f8b67739d07d7ac6a016ae16d6d679060c04613558adc5c"} Oct 04 11:42:20 crc kubenswrapper[5025]: I1004 11:42:20.432206 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39ce06265bcdf3687f8b67739d07d7ac6a016ae16d6d679060c04613558adc5c" Oct 04 11:42:20 crc kubenswrapper[5025]: I1004 11:42:20.478994 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:20 crc kubenswrapper[5025]: I1004 11:42:20.662811 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9qvd5"] Oct 04 11:42:22 crc kubenswrapper[5025]: I1004 11:42:22.450060 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9qvd5" podUID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerName="registry-server" containerID="cri-o://8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7" gracePeriod=2 Oct 04 11:42:22 crc kubenswrapper[5025]: I1004 11:42:22.887416 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.041681 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-catalog-content\") pod \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.041898 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wl5g5\" (UniqueName: \"kubernetes.io/projected/2bf77702-e805-4b2a-a8db-c8aa803c4e28-kube-api-access-wl5g5\") pod \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.042004 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-utilities\") pod \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\" (UID: \"2bf77702-e805-4b2a-a8db-c8aa803c4e28\") " Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.044261 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-utilities" (OuterVolumeSpecName: "utilities") pod "2bf77702-e805-4b2a-a8db-c8aa803c4e28" (UID: "2bf77702-e805-4b2a-a8db-c8aa803c4e28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.145213 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.330441 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bf77702-e805-4b2a-a8db-c8aa803c4e28-kube-api-access-wl5g5" (OuterVolumeSpecName: "kube-api-access-wl5g5") pod "2bf77702-e805-4b2a-a8db-c8aa803c4e28" (UID: "2bf77702-e805-4b2a-a8db-c8aa803c4e28"). InnerVolumeSpecName "kube-api-access-wl5g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.348639 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wl5g5\" (UniqueName: \"kubernetes.io/projected/2bf77702-e805-4b2a-a8db-c8aa803c4e28-kube-api-access-wl5g5\") on node \"crc\" DevicePath \"\"" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.435732 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bf77702-e805-4b2a-a8db-c8aa803c4e28" (UID: "2bf77702-e805-4b2a-a8db-c8aa803c4e28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.450064 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bf77702-e805-4b2a-a8db-c8aa803c4e28-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.460668 5025 generic.go:334] "Generic (PLEG): container finished" podID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerID="8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7" exitCode=0 Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.460717 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9qvd5" event={"ID":"2bf77702-e805-4b2a-a8db-c8aa803c4e28","Type":"ContainerDied","Data":"8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7"} Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.460743 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9qvd5" event={"ID":"2bf77702-e805-4b2a-a8db-c8aa803c4e28","Type":"ContainerDied","Data":"642e54c2d3955861b790b3da7e528252aff874641bf9ac4e4944e68defb17893"} Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.460759 5025 scope.go:117] "RemoveContainer" containerID="8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.460888 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9qvd5" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.502213 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9qvd5"] Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.502392 5025 scope.go:117] "RemoveContainer" containerID="2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.510734 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9qvd5"] Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.530637 5025 scope.go:117] "RemoveContainer" containerID="97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.580023 5025 scope.go:117] "RemoveContainer" containerID="8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7" Oct 04 11:42:23 crc kubenswrapper[5025]: E1004 11:42:23.585314 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7\": container with ID starting with 8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7 not found: ID does not exist" containerID="8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.585364 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7"} err="failed to get container status \"8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7\": rpc error: code = NotFound desc = could not find container \"8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7\": container with ID starting with 8c43660803726cefb5bf45110b2db5636471e1eb36346148a572f4f733859ac7 not found: ID does not exist" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.585395 5025 scope.go:117] "RemoveContainer" containerID="2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198" Oct 04 11:42:23 crc kubenswrapper[5025]: E1004 11:42:23.585972 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198\": container with ID starting with 2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198 not found: ID does not exist" containerID="2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.586068 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198"} err="failed to get container status \"2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198\": rpc error: code = NotFound desc = could not find container \"2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198\": container with ID starting with 2ddb981b5f9375c06f49cdba7f0b10646d8bf8b1d597f0fe9efe992d13caf198 not found: ID does not exist" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.586143 5025 scope.go:117] "RemoveContainer" containerID="97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c" Oct 04 11:42:23 crc kubenswrapper[5025]: E1004 11:42:23.586519 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c\": container with ID starting with 97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c not found: ID does not exist" containerID="97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.586746 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c"} err="failed to get container status \"97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c\": rpc error: code = NotFound desc = could not find container \"97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c\": container with ID starting with 97edc637e2019d90fd3b05b10a91f8f12e87db920dd55ae11c6839e9b1c5b79c not found: ID does not exist" Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.991662 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5vxlr/crc-debug-2rlhg"] Oct 04 11:42:23 crc kubenswrapper[5025]: I1004 11:42:23.998161 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5vxlr/crc-debug-2rlhg"] Oct 04 11:42:24 crc kubenswrapper[5025]: I1004 11:42:24.432717 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" path="/var/lib/kubelet/pods/2bf77702-e805-4b2a-a8db-c8aa803c4e28/volumes" Oct 04 11:42:24 crc kubenswrapper[5025]: I1004 11:42:24.434437 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47502e79-30b3-4af6-91fe-22658be4eb32" path="/var/lib/kubelet/pods/47502e79-30b3-4af6-91fe-22658be4eb32/volumes" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.156299 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-5vxlr/crc-debug-bvfnx"] Oct 04 11:42:25 crc kubenswrapper[5025]: E1004 11:42:25.156765 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerName="registry-server" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.156785 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerName="registry-server" Oct 04 11:42:25 crc kubenswrapper[5025]: E1004 11:42:25.156799 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47502e79-30b3-4af6-91fe-22658be4eb32" containerName="container-00" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.156806 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="47502e79-30b3-4af6-91fe-22658be4eb32" containerName="container-00" Oct 04 11:42:25 crc kubenswrapper[5025]: E1004 11:42:25.156815 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerName="extract-content" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.156822 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerName="extract-content" Oct 04 11:42:25 crc kubenswrapper[5025]: E1004 11:42:25.156832 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerName="extract-utilities" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.156838 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerName="extract-utilities" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.157003 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bf77702-e805-4b2a-a8db-c8aa803c4e28" containerName="registry-server" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.157033 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="47502e79-30b3-4af6-91fe-22658be4eb32" containerName="container-00" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.157709 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.284228 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-host\") pod \"crc-debug-bvfnx\" (UID: \"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb\") " pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.284436 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmvwc\" (UniqueName: \"kubernetes.io/projected/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-kube-api-access-tmvwc\") pod \"crc-debug-bvfnx\" (UID: \"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb\") " pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.386240 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-host\") pod \"crc-debug-bvfnx\" (UID: \"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb\") " pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.386381 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-host\") pod \"crc-debug-bvfnx\" (UID: \"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb\") " pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.386441 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmvwc\" (UniqueName: \"kubernetes.io/projected/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-kube-api-access-tmvwc\") pod \"crc-debug-bvfnx\" (UID: \"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb\") " pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.405458 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmvwc\" (UniqueName: \"kubernetes.io/projected/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-kube-api-access-tmvwc\") pod \"crc-debug-bvfnx\" (UID: \"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb\") " pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" Oct 04 11:42:25 crc kubenswrapper[5025]: I1004 11:42:25.489609 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" Oct 04 11:42:26 crc kubenswrapper[5025]: I1004 11:42:26.504466 5025 generic.go:334] "Generic (PLEG): container finished" podID="e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb" containerID="48b0749b52d51de05b48859d08e128577394a7f41e8b187a4089c3a37928a600" exitCode=0 Oct 04 11:42:26 crc kubenswrapper[5025]: I1004 11:42:26.504521 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" event={"ID":"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb","Type":"ContainerDied","Data":"48b0749b52d51de05b48859d08e128577394a7f41e8b187a4089c3a37928a600"} Oct 04 11:42:26 crc kubenswrapper[5025]: I1004 11:42:26.504545 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" event={"ID":"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb","Type":"ContainerStarted","Data":"b49d16a42993e624ad4e449480263da044efc479868e72f13efe64c54ce6fc4b"} Oct 04 11:42:26 crc kubenswrapper[5025]: I1004 11:42:26.545945 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5vxlr/crc-debug-bvfnx"] Oct 04 11:42:26 crc kubenswrapper[5025]: I1004 11:42:26.554151 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5vxlr/crc-debug-bvfnx"] Oct 04 11:42:27 crc kubenswrapper[5025]: I1004 11:42:27.411415 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:42:27 crc kubenswrapper[5025]: E1004 11:42:27.411665 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:42:27 crc kubenswrapper[5025]: I1004 11:42:27.640907 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" Oct 04 11:42:27 crc kubenswrapper[5025]: I1004 11:42:27.727844 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmvwc\" (UniqueName: \"kubernetes.io/projected/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-kube-api-access-tmvwc\") pod \"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb\" (UID: \"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb\") " Oct 04 11:42:27 crc kubenswrapper[5025]: I1004 11:42:27.728113 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-host\") pod \"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb\" (UID: \"e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb\") " Oct 04 11:42:27 crc kubenswrapper[5025]: I1004 11:42:27.728260 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-host" (OuterVolumeSpecName: "host") pod "e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb" (UID: "e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 04 11:42:27 crc kubenswrapper[5025]: I1004 11:42:27.728597 5025 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-host\") on node \"crc\" DevicePath \"\"" Oct 04 11:42:27 crc kubenswrapper[5025]: I1004 11:42:27.735520 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-kube-api-access-tmvwc" (OuterVolumeSpecName: "kube-api-access-tmvwc") pod "e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb" (UID: "e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb"). InnerVolumeSpecName "kube-api-access-tmvwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:42:27 crc kubenswrapper[5025]: I1004 11:42:27.830615 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmvwc\" (UniqueName: \"kubernetes.io/projected/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb-kube-api-access-tmvwc\") on node \"crc\" DevicePath \"\"" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.285296 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/util/0.log" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.423392 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb" path="/var/lib/kubelet/pods/e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb/volumes" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.467541 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/util/0.log" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.477684 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/pull/0.log" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.525976 5025 scope.go:117] "RemoveContainer" containerID="48b0749b52d51de05b48859d08e128577394a7f41e8b187a4089c3a37928a600" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.526126 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/crc-debug-bvfnx" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.535975 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/pull/0.log" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.715613 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/pull/0.log" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.723114 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/extract/0.log" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.729080 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_69a91769152834b0b8b99eda3b79a10374d31b97e919e850509435096astlb5_c5ba0046-11c4-4e7a-ae23-ff2759a29332/util/0.log" Oct 04 11:42:28 crc kubenswrapper[5025]: I1004 11:42:28.941479 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-58c4cd55f4-mr64p_91b206eb-240e-41b1-b1eb-df0c649554c9/kube-rbac-proxy/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.000718 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-58c4cd55f4-mr64p_91b206eb-240e-41b1-b1eb-df0c649554c9/manager/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.038844 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-dn4sj_b2f1ef79-fb7a-4d17-ad22-08a2f46df01f/kube-rbac-proxy/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.160794 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7d4d4f8d-dn4sj_b2f1ef79-fb7a-4d17-ad22-08a2f46df01f/manager/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.211450 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-ppf2w_0ed01e7a-3ef6-4cdd-97c5-ba85a84da593/kube-rbac-proxy/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.241542 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-75dfd9b554-ppf2w_0ed01e7a-3ef6-4cdd-97c5-ba85a84da593/manager/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.455292 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5dc44df7d5-tx2zp_b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228/kube-rbac-proxy/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.460739 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5dc44df7d5-tx2zp_b2c63d15-fe9b-4c3f-bf1a-ac61c49fa228/manager/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.560385 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-54b4974c45-hqbgk_b2b93be0-ff60-4432-b0db-2ec0e6c605a2/kube-rbac-proxy/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.607479 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-54b4974c45-hqbgk_b2b93be0-ff60-4432-b0db-2ec0e6c605a2/manager/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.675268 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-76d5b87f47-f2blx_6f5b4526-b51d-46e6-900d-492cf48c2710/kube-rbac-proxy/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.756651 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-76d5b87f47-f2blx_6f5b4526-b51d-46e6-900d-492cf48c2710/manager/0.log" Oct 04 11:42:29 crc kubenswrapper[5025]: I1004 11:42:29.799334 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-mjrsp_36628a66-3fd5-4ca5-a6e4-f6d59009b69f/kube-rbac-proxy/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.102256 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-f9rsb_8f554860-9fb5-4fc0-b795-832b03676469/kube-rbac-proxy/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.145562 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-658588b8c9-mjrsp_36628a66-3fd5-4ca5-a6e4-f6d59009b69f/manager/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.145608 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-649675d675-f9rsb_8f554860-9fb5-4fc0-b795-832b03676469/manager/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.285378 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b5ccf6d9c-ms9mz_bd496b6b-d68a-466b-a7de-a11157a5f470/kube-rbac-proxy/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.381853 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7b5ccf6d9c-ms9mz_bd496b6b-d68a-466b-a7de-a11157a5f470/manager/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.508474 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-r8csq_3ca07b7a-302b-474c-9aae-584e73350772/kube-rbac-proxy/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.565579 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-65d89cfd9f-r8csq_3ca07b7a-302b-474c-9aae-584e73350772/manager/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.594943 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx_b40c81d1-ba67-411d-b9ed-59a4991ac80f/kube-rbac-proxy/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.707426 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6cd6d7bdf5-9gwsx_b40c81d1-ba67-411d-b9ed-59a4991ac80f/manager/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.775693 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-fbp6d_dc8c4aca-63a0-490e-bf40-9377cdaefefc/kube-rbac-proxy/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.786066 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-8d984cc4d-fbp6d_dc8c4aca-63a0-490e-bf40-9377cdaefefc/manager/0.log" Oct 04 11:42:30 crc kubenswrapper[5025]: I1004 11:42:30.939471 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4svdm_b8cb1e86-b4dd-48e9-85c2-6c688d283b7d/kube-rbac-proxy/0.log" Oct 04 11:42:31 crc kubenswrapper[5025]: I1004 11:42:31.076243 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7c7fc454ff-4svdm_b8cb1e86-b4dd-48e9-85c2-6c688d283b7d/manager/0.log" Oct 04 11:42:31 crc kubenswrapper[5025]: I1004 11:42:31.153374 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-2llfv_e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc/manager/0.log" Oct 04 11:42:31 crc kubenswrapper[5025]: I1004 11:42:31.164520 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7468f855d8-2llfv_e84cb66c-9fa4-4b67-bb7d-e70a2f06ccfc/kube-rbac-proxy/0.log" Oct 04 11:42:31 crc kubenswrapper[5025]: I1004 11:42:31.278560 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx_2e70c89d-b3a2-4b15-90eb-91449857b0a2/kube-rbac-proxy/0.log" Oct 04 11:42:31 crc kubenswrapper[5025]: I1004 11:42:31.786391 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8c8588487-9l88t_745ffa84-a65e-48f0-aced-495c5be76951/kube-rbac-proxy/0.log" Oct 04 11:42:31 crc kubenswrapper[5025]: I1004 11:42:31.814097 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5dfbbd665ct6gjx_2e70c89d-b3a2-4b15-90eb-91449857b0a2/manager/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.060980 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-89c6d9c58-mtm9h_a1865149-283a-4933-b0d0-164afd0d1717/kube-rbac-proxy/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.147291 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-89c6d9c58-mtm9h_a1865149-283a-4933-b0d0-164afd0d1717/operator/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.155274 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-jjmsm_f3962134-5912-472b-903d-29f945e216e6/registry-server/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.309301 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6d8b6f9b9-dtnvw_0371bb2b-35e9-4770-9fc4-560d5c2c343f/kube-rbac-proxy/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.452364 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6d8b6f9b9-dtnvw_0371bb2b-35e9-4770-9fc4-560d5c2c343f/manager/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.589989 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-f6bvd_81158795-65c7-4a01-b9ba-ad40b7d22582/kube-rbac-proxy/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.610511 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-54689d9f88-f6bvd_81158795-65c7-4a01-b9ba-ad40b7d22582/manager/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.648712 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-9c4ml_a48ece71-fc88-4329-8ad3-fe9db58bb99a/operator/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.884894 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-8c8588487-9l88t_745ffa84-a65e-48f0-aced-495c5be76951/manager/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.899584 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-cl2kh_ad4743c5-3d51-421c-98b8-1463831fb92a/manager/0.log" Oct 04 11:42:32 crc kubenswrapper[5025]: I1004 11:42:32.913753 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6859f9b676-cl2kh_ad4743c5-3d51-421c-98b8-1463831fb92a/kube-rbac-proxy/0.log" Oct 04 11:42:33 crc kubenswrapper[5025]: I1004 11:42:33.797476 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-jrnhn_85d359e6-3627-4c6b-b974-3599728017a3/kube-rbac-proxy/0.log" Oct 04 11:42:33 crc kubenswrapper[5025]: I1004 11:42:33.984506 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-zmt6f_a1cffbcd-c643-4b18-8832-d28767b38fc6/kube-rbac-proxy/0.log" Oct 04 11:42:34 crc kubenswrapper[5025]: I1004 11:42:34.003572 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-5d4d74dd89-jrnhn_85d359e6-3627-4c6b-b974-3599728017a3/manager/0.log" Oct 04 11:42:34 crc kubenswrapper[5025]: I1004 11:42:34.003661 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cd5cb47d7-zmt6f_a1cffbcd-c643-4b18-8832-d28767b38fc6/manager/0.log" Oct 04 11:42:34 crc kubenswrapper[5025]: I1004 11:42:34.284764 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-lfgrl_80fb065c-ce02-4189-8b7a-7bbe78181ee9/kube-rbac-proxy/0.log" Oct 04 11:42:34 crc kubenswrapper[5025]: I1004 11:42:34.353055 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6cbc6dd547-lfgrl_80fb065c-ce02-4189-8b7a-7bbe78181ee9/manager/0.log" Oct 04 11:42:39 crc kubenswrapper[5025]: I1004 11:42:39.411153 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:42:39 crc kubenswrapper[5025]: E1004 11:42:39.411747 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:42:50 crc kubenswrapper[5025]: I1004 11:42:50.612905 5025 scope.go:117] "RemoveContainer" containerID="f0cc21280f85ac59b6c3950d17b7b815955adce30405f23d34162c183ae336c9" Oct 04 11:42:51 crc kubenswrapper[5025]: I1004 11:42:51.329381 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-5k929_4f79bb2f-4697-4f2f-80f5-c95f59b3273f/control-plane-machine-set-operator/0.log" Oct 04 11:42:51 crc kubenswrapper[5025]: I1004 11:42:51.490833 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6mdn4_227aafde-cb49-4cff-b561-d7ba6b84a1cc/machine-api-operator/0.log" Oct 04 11:42:51 crc kubenswrapper[5025]: I1004 11:42:51.502747 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-6mdn4_227aafde-cb49-4cff-b561-d7ba6b84a1cc/kube-rbac-proxy/0.log" Oct 04 11:42:53 crc kubenswrapper[5025]: I1004 11:42:53.412652 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:42:53 crc kubenswrapper[5025]: E1004 11:42:53.414111 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:43:04 crc kubenswrapper[5025]: I1004 11:43:04.331505 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-hxgr5_8f18ef59-714b-4044-b76f-44cb4b523a3f/cert-manager-controller/0.log" Oct 04 11:43:04 crc kubenswrapper[5025]: I1004 11:43:04.396059 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-54hps_a413e9a9-6200-4e4a-abb7-939ed023a70f/cert-manager-cainjector/0.log" Oct 04 11:43:04 crc kubenswrapper[5025]: I1004 11:43:04.555725 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-8h2xb_a45d9552-c070-4b0b-99b7-50cfd7acbd28/cert-manager-webhook/0.log" Oct 04 11:43:08 crc kubenswrapper[5025]: I1004 11:43:08.412508 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:43:08 crc kubenswrapper[5025]: E1004 11:43:08.413557 5025 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-2dll9_openshift-machine-config-operator(54919b0d-887d-4727-adfc-e48a66e680ba)\"" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" Oct 04 11:43:16 crc kubenswrapper[5025]: I1004 11:43:16.497105 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-mtz64_56180570-2df4-4eee-93dd-d403e7c4a24f/nmstate-console-plugin/0.log" Oct 04 11:43:16 crc kubenswrapper[5025]: I1004 11:43:16.691078 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-ksclx_1c998df0-be41-4019-bba6-5e1b4f58a10e/nmstate-handler/0.log" Oct 04 11:43:16 crc kubenswrapper[5025]: I1004 11:43:16.733232 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-55hjr_7604348c-0bc1-4932-8ff6-00b3098c7326/kube-rbac-proxy/0.log" Oct 04 11:43:16 crc kubenswrapper[5025]: I1004 11:43:16.757027 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-55hjr_7604348c-0bc1-4932-8ff6-00b3098c7326/nmstate-metrics/0.log" Oct 04 11:43:16 crc kubenswrapper[5025]: I1004 11:43:16.921388 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-dvdx6_3b47ddfd-e0dd-4a6f-9879-505058ddc2b3/nmstate-operator/0.log" Oct 04 11:43:16 crc kubenswrapper[5025]: I1004 11:43:16.973513 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-g2ggs_4f852867-9d9b-4255-898c-3b26aabf3243/nmstate-webhook/0.log" Oct 04 11:43:23 crc kubenswrapper[5025]: I1004 11:43:23.411632 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:43:24 crc kubenswrapper[5025]: I1004 11:43:24.116774 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"9ee7c3d769c3511331938367ebe24b4dff8cf66c7554d7ecd518bff528e5c87b"} Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.074093 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-jqv9k_24bfff08-7460-44c2-8358-c18faa7933fd/kube-rbac-proxy/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.179533 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-jqv9k_24bfff08-7460-44c2-8358-c18faa7933fd/controller/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.311489 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-frr-files/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.491674 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-frr-files/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.510979 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-metrics/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.552083 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-reloader/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.565247 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-reloader/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.761520 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-metrics/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.762054 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-reloader/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.833840 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-frr-files/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.838363 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-metrics/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.975642 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-frr-files/0.log" Oct 04 11:43:31 crc kubenswrapper[5025]: I1004 11:43:31.987126 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-reloader/0.log" Oct 04 11:43:32 crc kubenswrapper[5025]: I1004 11:43:32.044954 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/controller/0.log" Oct 04 11:43:32 crc kubenswrapper[5025]: I1004 11:43:32.075870 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/cp-metrics/0.log" Oct 04 11:43:32 crc kubenswrapper[5025]: I1004 11:43:32.216247 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/frr-metrics/0.log" Oct 04 11:43:32 crc kubenswrapper[5025]: I1004 11:43:32.265188 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/kube-rbac-proxy/0.log" Oct 04 11:43:32 crc kubenswrapper[5025]: I1004 11:43:32.276382 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/kube-rbac-proxy-frr/0.log" Oct 04 11:43:32 crc kubenswrapper[5025]: I1004 11:43:32.433839 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/reloader/0.log" Oct 04 11:43:32 crc kubenswrapper[5025]: I1004 11:43:32.504426 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-ptxm6_27b1b14d-46e4-45e2-8bcf-b37d8b0a7f79/frr-k8s-webhook-server/0.log" Oct 04 11:43:33 crc kubenswrapper[5025]: I1004 11:43:33.400696 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-66f54ffcff-9lmgs_b87c8f6a-58fe-47e2-822b-ad21c79d0123/manager/0.log" Oct 04 11:43:33 crc kubenswrapper[5025]: I1004 11:43:33.452351 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-764d6c8955-dhmw9_20275cb8-c871-4219-8aee-4f47f917ed72/webhook-server/0.log" Oct 04 11:43:33 crc kubenswrapper[5025]: I1004 11:43:33.688755 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6rf4m_d9c990ef-7b47-4ece-b38d-822182cd7407/kube-rbac-proxy/0.log" Oct 04 11:43:33 crc kubenswrapper[5025]: I1004 11:43:33.720077 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-7rrhp_3aa58e9d-ce3d-4f94-aec2-aa1d870d309b/frr/0.log" Oct 04 11:43:33 crc kubenswrapper[5025]: I1004 11:43:33.995961 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6rf4m_d9c990ef-7b47-4ece-b38d-822182cd7407/speaker/0.log" Oct 04 11:43:47 crc kubenswrapper[5025]: I1004 11:43:47.735892 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/util/0.log" Oct 04 11:43:47 crc kubenswrapper[5025]: I1004 11:43:47.897128 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/util/0.log" Oct 04 11:43:47 crc kubenswrapper[5025]: I1004 11:43:47.926173 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/pull/0.log" Oct 04 11:43:47 crc kubenswrapper[5025]: I1004 11:43:47.953336 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/pull/0.log" Oct 04 11:43:48 crc kubenswrapper[5025]: I1004 11:43:48.073145 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/pull/0.log" Oct 04 11:43:48 crc kubenswrapper[5025]: I1004 11:43:48.082677 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/util/0.log" Oct 04 11:43:48 crc kubenswrapper[5025]: I1004 11:43:48.097723 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2fr5cb_6257e369-9d92-403e-85cf-3990b895da8d/extract/0.log" Oct 04 11:43:48 crc kubenswrapper[5025]: I1004 11:43:48.244246 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-utilities/0.log" Oct 04 11:43:48 crc kubenswrapper[5025]: I1004 11:43:48.414419 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-utilities/0.log" Oct 04 11:43:48 crc kubenswrapper[5025]: I1004 11:43:48.427416 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-content/0.log" Oct 04 11:43:48 crc kubenswrapper[5025]: I1004 11:43:48.458753 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-content/0.log" Oct 04 11:43:48 crc kubenswrapper[5025]: I1004 11:43:48.601995 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-content/0.log" Oct 04 11:43:48 crc kubenswrapper[5025]: I1004 11:43:48.620742 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/extract-utilities/0.log" Oct 04 11:43:48 crc kubenswrapper[5025]: I1004 11:43:48.851159 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-utilities/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.078422 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-utilities/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.080625 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5ln4x_ecc4c1d1-a9f7-4cf4-b1d7-5336507676c5/registry-server/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.091559 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-content/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.110398 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-content/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.280692 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-utilities/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.305883 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/extract-content/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.461347 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/util/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.662650 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/util/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.711685 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/pull/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.732200 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/pull/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.802854 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-d2hs7_a4a6c356-d827-4b99-a7f2-fbfba013450b/registry-server/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.930120 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/util/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.947461 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/pull/0.log" Oct 04 11:43:49 crc kubenswrapper[5025]: I1004 11:43:49.953514 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cg95hr_33dbe626-f29f-4939-bcdc-c04109e9820c/extract/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.119066 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-utilities/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.136396 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-d2str_894ba87a-0076-47c2-836d-c972664c8900/marketplace-operator/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.275569 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-content/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.323002 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-utilities/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.334967 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-content/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.472715 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-utilities/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.479246 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/extract-content/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.580128 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-utilities/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.611548 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-m9hl2_c3e7faef-f9a9-4742-b6e1-893e70ab32b3/registry-server/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.756541 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-content/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.761635 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-content/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.779052 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-utilities/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.938498 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-content/0.log" Oct 04 11:43:50 crc kubenswrapper[5025]: I1004 11:43:50.952552 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/extract-utilities/0.log" Oct 04 11:43:51 crc kubenswrapper[5025]: I1004 11:43:51.610339 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-rxpjg_1b1dd0c3-486f-4ac8-be1b-cfc100de913e/registry-server/0.log" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.277772 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-l695n"] Oct 04 11:44:32 crc kubenswrapper[5025]: E1004 11:44:32.278844 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb" containerName="container-00" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.278859 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb" containerName="container-00" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.316487 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2b4ce9e-e629-47c2-81c3-2c58cdd64ceb" containerName="container-00" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.328378 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.348729 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l695n"] Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.494619 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-utilities\") pod \"community-operators-l695n\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.495147 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-catalog-content\") pod \"community-operators-l695n\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.495295 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swtms\" (UniqueName: \"kubernetes.io/projected/eb0c7370-fa61-47b9-9def-cb16e6692379-kube-api-access-swtms\") pod \"community-operators-l695n\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.597368 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-catalog-content\") pod \"community-operators-l695n\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.597466 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swtms\" (UniqueName: \"kubernetes.io/projected/eb0c7370-fa61-47b9-9def-cb16e6692379-kube-api-access-swtms\") pod \"community-operators-l695n\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.597545 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-utilities\") pod \"community-operators-l695n\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.598049 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-catalog-content\") pod \"community-operators-l695n\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.598116 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-utilities\") pod \"community-operators-l695n\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.637418 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swtms\" (UniqueName: \"kubernetes.io/projected/eb0c7370-fa61-47b9-9def-cb16e6692379-kube-api-access-swtms\") pod \"community-operators-l695n\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:32 crc kubenswrapper[5025]: I1004 11:44:32.676951 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:33 crc kubenswrapper[5025]: I1004 11:44:33.254796 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-l695n"] Oct 04 11:44:33 crc kubenswrapper[5025]: I1004 11:44:33.796346 5025 generic.go:334] "Generic (PLEG): container finished" podID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerID="471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3" exitCode=0 Oct 04 11:44:33 crc kubenswrapper[5025]: I1004 11:44:33.796685 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l695n" event={"ID":"eb0c7370-fa61-47b9-9def-cb16e6692379","Type":"ContainerDied","Data":"471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3"} Oct 04 11:44:33 crc kubenswrapper[5025]: I1004 11:44:33.796730 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l695n" event={"ID":"eb0c7370-fa61-47b9-9def-cb16e6692379","Type":"ContainerStarted","Data":"7b6cebfb2a117bec86d1fa68267e59c65b3433bd37cf436e4123927f78b1d1f9"} Oct 04 11:44:33 crc kubenswrapper[5025]: I1004 11:44:33.798533 5025 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 04 11:44:35 crc kubenswrapper[5025]: I1004 11:44:35.822140 5025 generic.go:334] "Generic (PLEG): container finished" podID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerID="963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c" exitCode=0 Oct 04 11:44:35 crc kubenswrapper[5025]: I1004 11:44:35.822255 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l695n" event={"ID":"eb0c7370-fa61-47b9-9def-cb16e6692379","Type":"ContainerDied","Data":"963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c"} Oct 04 11:44:36 crc kubenswrapper[5025]: I1004 11:44:36.833446 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l695n" event={"ID":"eb0c7370-fa61-47b9-9def-cb16e6692379","Type":"ContainerStarted","Data":"0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c"} Oct 04 11:44:36 crc kubenswrapper[5025]: I1004 11:44:36.855032 5025 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-l695n" podStartSLOduration=2.415008627 podStartE2EDuration="4.855005404s" podCreationTimestamp="2025-10-04 11:44:32 +0000 UTC" firstStartedPulling="2025-10-04 11:44:33.798272007 +0000 UTC m=+4202.223238887" lastFinishedPulling="2025-10-04 11:44:36.238268754 +0000 UTC m=+4204.663235664" observedRunningTime="2025-10-04 11:44:36.851392592 +0000 UTC m=+4205.276359472" watchObservedRunningTime="2025-10-04 11:44:36.855005404 +0000 UTC m=+4205.279972284" Oct 04 11:44:42 crc kubenswrapper[5025]: I1004 11:44:42.677907 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:42 crc kubenswrapper[5025]: I1004 11:44:42.678532 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:42 crc kubenswrapper[5025]: I1004 11:44:42.753471 5025 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:42 crc kubenswrapper[5025]: I1004 11:44:42.969339 5025 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:43 crc kubenswrapper[5025]: I1004 11:44:43.026188 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l695n"] Oct 04 11:44:44 crc kubenswrapper[5025]: I1004 11:44:44.925321 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-l695n" podUID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerName="registry-server" containerID="cri-o://0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c" gracePeriod=2 Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.392270 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.559685 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-utilities\") pod \"eb0c7370-fa61-47b9-9def-cb16e6692379\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.559780 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swtms\" (UniqueName: \"kubernetes.io/projected/eb0c7370-fa61-47b9-9def-cb16e6692379-kube-api-access-swtms\") pod \"eb0c7370-fa61-47b9-9def-cb16e6692379\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.559881 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-catalog-content\") pod \"eb0c7370-fa61-47b9-9def-cb16e6692379\" (UID: \"eb0c7370-fa61-47b9-9def-cb16e6692379\") " Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.561163 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-utilities" (OuterVolumeSpecName: "utilities") pod "eb0c7370-fa61-47b9-9def-cb16e6692379" (UID: "eb0c7370-fa61-47b9-9def-cb16e6692379"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.568395 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb0c7370-fa61-47b9-9def-cb16e6692379-kube-api-access-swtms" (OuterVolumeSpecName: "kube-api-access-swtms") pod "eb0c7370-fa61-47b9-9def-cb16e6692379" (UID: "eb0c7370-fa61-47b9-9def-cb16e6692379"). InnerVolumeSpecName "kube-api-access-swtms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.635477 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb0c7370-fa61-47b9-9def-cb16e6692379" (UID: "eb0c7370-fa61-47b9-9def-cb16e6692379"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.661899 5025 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-utilities\") on node \"crc\" DevicePath \"\"" Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.661928 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swtms\" (UniqueName: \"kubernetes.io/projected/eb0c7370-fa61-47b9-9def-cb16e6692379-kube-api-access-swtms\") on node \"crc\" DevicePath \"\"" Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.661944 5025 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb0c7370-fa61-47b9-9def-cb16e6692379-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.940030 5025 generic.go:334] "Generic (PLEG): container finished" podID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerID="0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c" exitCode=0 Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.940083 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l695n" event={"ID":"eb0c7370-fa61-47b9-9def-cb16e6692379","Type":"ContainerDied","Data":"0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c"} Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.940121 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-l695n" Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.940135 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-l695n" event={"ID":"eb0c7370-fa61-47b9-9def-cb16e6692379","Type":"ContainerDied","Data":"7b6cebfb2a117bec86d1fa68267e59c65b3433bd37cf436e4123927f78b1d1f9"} Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.940162 5025 scope.go:117] "RemoveContainer" containerID="0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c" Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.989222 5025 scope.go:117] "RemoveContainer" containerID="963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c" Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.989856 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-l695n"] Oct 04 11:44:45 crc kubenswrapper[5025]: I1004 11:44:45.997100 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-l695n"] Oct 04 11:44:46 crc kubenswrapper[5025]: I1004 11:44:46.014797 5025 scope.go:117] "RemoveContainer" containerID="471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3" Oct 04 11:44:46 crc kubenswrapper[5025]: I1004 11:44:46.070380 5025 scope.go:117] "RemoveContainer" containerID="0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c" Oct 04 11:44:46 crc kubenswrapper[5025]: E1004 11:44:46.070983 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c\": container with ID starting with 0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c not found: ID does not exist" containerID="0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c" Oct 04 11:44:46 crc kubenswrapper[5025]: I1004 11:44:46.071190 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c"} err="failed to get container status \"0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c\": rpc error: code = NotFound desc = could not find container \"0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c\": container with ID starting with 0783dd0b66d6758389b68fa4aa04413bf110d66d417a629ec26fa24c9302cb4c not found: ID does not exist" Oct 04 11:44:46 crc kubenswrapper[5025]: I1004 11:44:46.071350 5025 scope.go:117] "RemoveContainer" containerID="963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c" Oct 04 11:44:46 crc kubenswrapper[5025]: E1004 11:44:46.071969 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c\": container with ID starting with 963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c not found: ID does not exist" containerID="963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c" Oct 04 11:44:46 crc kubenswrapper[5025]: I1004 11:44:46.072007 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c"} err="failed to get container status \"963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c\": rpc error: code = NotFound desc = could not find container \"963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c\": container with ID starting with 963fa55f726557efbb4c3703c3097c976327b6798be6c4267aa6587bde70810c not found: ID does not exist" Oct 04 11:44:46 crc kubenswrapper[5025]: I1004 11:44:46.072044 5025 scope.go:117] "RemoveContainer" containerID="471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3" Oct 04 11:44:46 crc kubenswrapper[5025]: E1004 11:44:46.072300 5025 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3\": container with ID starting with 471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3 not found: ID does not exist" containerID="471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3" Oct 04 11:44:46 crc kubenswrapper[5025]: I1004 11:44:46.072333 5025 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3"} err="failed to get container status \"471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3\": rpc error: code = NotFound desc = could not find container \"471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3\": container with ID starting with 471b92dabfab06a2f66648b2823e2e2f475c365372d91b8850161c87ac5883a3 not found: ID does not exist" Oct 04 11:44:46 crc kubenswrapper[5025]: I1004 11:44:46.434620 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb0c7370-fa61-47b9-9def-cb16e6692379" path="/var/lib/kubelet/pods/eb0c7370-fa61-47b9-9def-cb16e6692379/volumes" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.164025 5025 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs"] Oct 04 11:45:00 crc kubenswrapper[5025]: E1004 11:45:00.164902 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerName="registry-server" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.164915 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerName="registry-server" Oct 04 11:45:00 crc kubenswrapper[5025]: E1004 11:45:00.164942 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerName="extract-utilities" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.164949 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerName="extract-utilities" Oct 04 11:45:00 crc kubenswrapper[5025]: E1004 11:45:00.164969 5025 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerName="extract-content" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.164977 5025 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerName="extract-content" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.165189 5025 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb0c7370-fa61-47b9-9def-cb16e6692379" containerName="registry-server" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.165823 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.167765 5025 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.168595 5025 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.194569 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs"] Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.298110 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-config-volume\") pod \"collect-profiles-29326305-l9djs\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.298168 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg96j\" (UniqueName: \"kubernetes.io/projected/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-kube-api-access-mg96j\") pod \"collect-profiles-29326305-l9djs\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.298240 5025 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-secret-volume\") pod \"collect-profiles-29326305-l9djs\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.400383 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-config-volume\") pod \"collect-profiles-29326305-l9djs\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.400480 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg96j\" (UniqueName: \"kubernetes.io/projected/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-kube-api-access-mg96j\") pod \"collect-profiles-29326305-l9djs\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.400620 5025 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-secret-volume\") pod \"collect-profiles-29326305-l9djs\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.401954 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-config-volume\") pod \"collect-profiles-29326305-l9djs\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.411992 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-secret-volume\") pod \"collect-profiles-29326305-l9djs\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.432390 5025 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg96j\" (UniqueName: \"kubernetes.io/projected/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-kube-api-access-mg96j\") pod \"collect-profiles-29326305-l9djs\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.491765 5025 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:00 crc kubenswrapper[5025]: I1004 11:45:00.976850 5025 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs"] Oct 04 11:45:02 crc kubenswrapper[5025]: I1004 11:45:02.147317 5025 generic.go:334] "Generic (PLEG): container finished" podID="e2b8f0d8-f3b3-4618-8913-952a777c3d2d" containerID="60a2aa4106db6719d7f761f72aa357aca342c4165cfeea4655ede4218a9b5a9f" exitCode=0 Oct 04 11:45:02 crc kubenswrapper[5025]: I1004 11:45:02.147472 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" event={"ID":"e2b8f0d8-f3b3-4618-8913-952a777c3d2d","Type":"ContainerDied","Data":"60a2aa4106db6719d7f761f72aa357aca342c4165cfeea4655ede4218a9b5a9f"} Oct 04 11:45:02 crc kubenswrapper[5025]: I1004 11:45:02.147841 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" event={"ID":"e2b8f0d8-f3b3-4618-8913-952a777c3d2d","Type":"ContainerStarted","Data":"3296814af4f3cee11cdf17aafdf8228939736d90f7f1f88d88153540ef8d56a7"} Oct 04 11:45:03 crc kubenswrapper[5025]: I1004 11:45:03.483445 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:03 crc kubenswrapper[5025]: I1004 11:45:03.560642 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-secret-volume\") pod \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " Oct 04 11:45:03 crc kubenswrapper[5025]: I1004 11:45:03.560703 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-config-volume\") pod \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " Oct 04 11:45:03 crc kubenswrapper[5025]: I1004 11:45:03.560741 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg96j\" (UniqueName: \"kubernetes.io/projected/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-kube-api-access-mg96j\") pod \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\" (UID: \"e2b8f0d8-f3b3-4618-8913-952a777c3d2d\") " Oct 04 11:45:03 crc kubenswrapper[5025]: I1004 11:45:03.561874 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-config-volume" (OuterVolumeSpecName: "config-volume") pod "e2b8f0d8-f3b3-4618-8913-952a777c3d2d" (UID: "e2b8f0d8-f3b3-4618-8913-952a777c3d2d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 04 11:45:03 crc kubenswrapper[5025]: I1004 11:45:03.611235 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-kube-api-access-mg96j" (OuterVolumeSpecName: "kube-api-access-mg96j") pod "e2b8f0d8-f3b3-4618-8913-952a777c3d2d" (UID: "e2b8f0d8-f3b3-4618-8913-952a777c3d2d"). InnerVolumeSpecName "kube-api-access-mg96j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:45:03 crc kubenswrapper[5025]: I1004 11:45:03.611425 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e2b8f0d8-f3b3-4618-8913-952a777c3d2d" (UID: "e2b8f0d8-f3b3-4618-8913-952a777c3d2d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 04 11:45:03 crc kubenswrapper[5025]: I1004 11:45:03.663315 5025 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:45:03 crc kubenswrapper[5025]: I1004 11:45:03.663358 5025 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 04 11:45:03 crc kubenswrapper[5025]: I1004 11:45:03.663373 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg96j\" (UniqueName: \"kubernetes.io/projected/e2b8f0d8-f3b3-4618-8913-952a777c3d2d-kube-api-access-mg96j\") on node \"crc\" DevicePath \"\"" Oct 04 11:45:04 crc kubenswrapper[5025]: I1004 11:45:04.181161 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" event={"ID":"e2b8f0d8-f3b3-4618-8913-952a777c3d2d","Type":"ContainerDied","Data":"3296814af4f3cee11cdf17aafdf8228939736d90f7f1f88d88153540ef8d56a7"} Oct 04 11:45:04 crc kubenswrapper[5025]: I1004 11:45:04.181229 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3296814af4f3cee11cdf17aafdf8228939736d90f7f1f88d88153540ef8d56a7" Oct 04 11:45:04 crc kubenswrapper[5025]: I1004 11:45:04.181228 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29326305-l9djs" Oct 04 11:45:04 crc kubenswrapper[5025]: I1004 11:45:04.579182 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq"] Oct 04 11:45:04 crc kubenswrapper[5025]: I1004 11:45:04.587173 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29326260-pmfcq"] Oct 04 11:45:06 crc kubenswrapper[5025]: I1004 11:45:06.433358 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93845871-6c40-43cc-b144-257d6e4385af" path="/var/lib/kubelet/pods/93845871-6c40-43cc-b144-257d6e4385af/volumes" Oct 04 11:45:44 crc kubenswrapper[5025]: I1004 11:45:44.713437 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:45:44 crc kubenswrapper[5025]: I1004 11:45:44.713970 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:45:50 crc kubenswrapper[5025]: I1004 11:45:50.692158 5025 generic.go:334] "Generic (PLEG): container finished" podID="c51b4c7e-c4de-4022-9965-0cc6c67bc31e" containerID="9c120ab46ead71029270acaf51a8b5336e6feaaed740c204ac8ab3256e113c44" exitCode=0 Oct 04 11:45:50 crc kubenswrapper[5025]: I1004 11:45:50.692272 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-5vxlr/must-gather-r79tc" event={"ID":"c51b4c7e-c4de-4022-9965-0cc6c67bc31e","Type":"ContainerDied","Data":"9c120ab46ead71029270acaf51a8b5336e6feaaed740c204ac8ab3256e113c44"} Oct 04 11:45:50 crc kubenswrapper[5025]: I1004 11:45:50.693404 5025 scope.go:117] "RemoveContainer" containerID="9c120ab46ead71029270acaf51a8b5336e6feaaed740c204ac8ab3256e113c44" Oct 04 11:45:50 crc kubenswrapper[5025]: I1004 11:45:50.797280 5025 scope.go:117] "RemoveContainer" containerID="dbe57f783eb1381a271a6a4e58f90b594bf318a37cd0994a56eb904f52e7cc4d" Oct 04 11:45:51 crc kubenswrapper[5025]: I1004 11:45:51.513367 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5vxlr_must-gather-r79tc_c51b4c7e-c4de-4022-9965-0cc6c67bc31e/gather/0.log" Oct 04 11:46:03 crc kubenswrapper[5025]: I1004 11:46:03.407299 5025 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-5vxlr/must-gather-r79tc"] Oct 04 11:46:03 crc kubenswrapper[5025]: I1004 11:46:03.408035 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-5vxlr/must-gather-r79tc" podUID="c51b4c7e-c4de-4022-9965-0cc6c67bc31e" containerName="copy" containerID="cri-o://7ee702bbe93c8df4806c249c16e94cf87ad779b7ee37be0028d6057fbe4f4d56" gracePeriod=2 Oct 04 11:46:03 crc kubenswrapper[5025]: I1004 11:46:03.427953 5025 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-5vxlr/must-gather-r79tc"] Oct 04 11:46:03 crc kubenswrapper[5025]: I1004 11:46:03.819277 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5vxlr_must-gather-r79tc_c51b4c7e-c4de-4022-9965-0cc6c67bc31e/copy/0.log" Oct 04 11:46:03 crc kubenswrapper[5025]: I1004 11:46:03.819979 5025 generic.go:334] "Generic (PLEG): container finished" podID="c51b4c7e-c4de-4022-9965-0cc6c67bc31e" containerID="7ee702bbe93c8df4806c249c16e94cf87ad779b7ee37be0028d6057fbe4f4d56" exitCode=143 Oct 04 11:46:03 crc kubenswrapper[5025]: I1004 11:46:03.820039 5025 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffc968388e1105be6cd54f8afd4e94ca03d044ddce1c9b32ba863f65128c2ad5" Oct 04 11:46:03 crc kubenswrapper[5025]: I1004 11:46:03.880622 5025 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-5vxlr_must-gather-r79tc_c51b4c7e-c4de-4022-9965-0cc6c67bc31e/copy/0.log" Oct 04 11:46:03 crc kubenswrapper[5025]: I1004 11:46:03.881058 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/must-gather-r79tc" Oct 04 11:46:04 crc kubenswrapper[5025]: I1004 11:46:04.002959 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-must-gather-output\") pod \"c51b4c7e-c4de-4022-9965-0cc6c67bc31e\" (UID: \"c51b4c7e-c4de-4022-9965-0cc6c67bc31e\") " Oct 04 11:46:04 crc kubenswrapper[5025]: I1004 11:46:04.003046 5025 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzqjj\" (UniqueName: \"kubernetes.io/projected/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-kube-api-access-dzqjj\") pod \"c51b4c7e-c4de-4022-9965-0cc6c67bc31e\" (UID: \"c51b4c7e-c4de-4022-9965-0cc6c67bc31e\") " Oct 04 11:46:04 crc kubenswrapper[5025]: I1004 11:46:04.010969 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-kube-api-access-dzqjj" (OuterVolumeSpecName: "kube-api-access-dzqjj") pod "c51b4c7e-c4de-4022-9965-0cc6c67bc31e" (UID: "c51b4c7e-c4de-4022-9965-0cc6c67bc31e"). InnerVolumeSpecName "kube-api-access-dzqjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 04 11:46:04 crc kubenswrapper[5025]: I1004 11:46:04.105065 5025 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzqjj\" (UniqueName: \"kubernetes.io/projected/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-kube-api-access-dzqjj\") on node \"crc\" DevicePath \"\"" Oct 04 11:46:04 crc kubenswrapper[5025]: I1004 11:46:04.159494 5025 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c51b4c7e-c4de-4022-9965-0cc6c67bc31e" (UID: "c51b4c7e-c4de-4022-9965-0cc6c67bc31e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 04 11:46:04 crc kubenswrapper[5025]: I1004 11:46:04.207259 5025 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c51b4c7e-c4de-4022-9965-0cc6c67bc31e-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 04 11:46:04 crc kubenswrapper[5025]: I1004 11:46:04.430862 5025 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c51b4c7e-c4de-4022-9965-0cc6c67bc31e" path="/var/lib/kubelet/pods/c51b4c7e-c4de-4022-9965-0cc6c67bc31e/volumes" Oct 04 11:46:04 crc kubenswrapper[5025]: I1004 11:46:04.828688 5025 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-5vxlr/must-gather-r79tc" Oct 04 11:46:14 crc kubenswrapper[5025]: I1004 11:46:14.714095 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:46:14 crc kubenswrapper[5025]: I1004 11:46:14.714698 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:46:44 crc kubenswrapper[5025]: I1004 11:46:44.713213 5025 patch_prober.go:28] interesting pod/machine-config-daemon-2dll9 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 04 11:46:44 crc kubenswrapper[5025]: I1004 11:46:44.713693 5025 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 04 11:46:44 crc kubenswrapper[5025]: I1004 11:46:44.713737 5025 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" Oct 04 11:46:44 crc kubenswrapper[5025]: I1004 11:46:44.714545 5025 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ee7c3d769c3511331938367ebe24b4dff8cf66c7554d7ecd518bff528e5c87b"} pod="openshift-machine-config-operator/machine-config-daemon-2dll9" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 04 11:46:44 crc kubenswrapper[5025]: I1004 11:46:44.714599 5025 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" podUID="54919b0d-887d-4727-adfc-e48a66e680ba" containerName="machine-config-daemon" containerID="cri-o://9ee7c3d769c3511331938367ebe24b4dff8cf66c7554d7ecd518bff528e5c87b" gracePeriod=600 Oct 04 11:46:45 crc kubenswrapper[5025]: I1004 11:46:45.238813 5025 generic.go:334] "Generic (PLEG): container finished" podID="54919b0d-887d-4727-adfc-e48a66e680ba" containerID="9ee7c3d769c3511331938367ebe24b4dff8cf66c7554d7ecd518bff528e5c87b" exitCode=0 Oct 04 11:46:45 crc kubenswrapper[5025]: I1004 11:46:45.238903 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerDied","Data":"9ee7c3d769c3511331938367ebe24b4dff8cf66c7554d7ecd518bff528e5c87b"} Oct 04 11:46:45 crc kubenswrapper[5025]: I1004 11:46:45.238974 5025 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-2dll9" event={"ID":"54919b0d-887d-4727-adfc-e48a66e680ba","Type":"ContainerStarted","Data":"e9c005e9bd205f756c6dfb848d75bccbf739c8501ed3b3241add3995de9dd26e"} Oct 04 11:46:45 crc kubenswrapper[5025]: I1004 11:46:45.239000 5025 scope.go:117] "RemoveContainer" containerID="90655883eb81f0c5594ea998d1936d9aa5f642337dd1a80d01918a9d51764d51" Oct 04 11:46:50 crc kubenswrapper[5025]: I1004 11:46:50.888163 5025 scope.go:117] "RemoveContainer" containerID="9c120ab46ead71029270acaf51a8b5336e6feaaed740c204ac8ab3256e113c44" Oct 04 11:46:51 crc kubenswrapper[5025]: I1004 11:46:51.002150 5025 scope.go:117] "RemoveContainer" containerID="d56e864224e31d091d0d1d07fd68b9db2c60024d7b9e883e39745d927b71cc4b" Oct 04 11:46:51 crc kubenswrapper[5025]: I1004 11:46:51.024962 5025 scope.go:117] "RemoveContainer" containerID="7ee702bbe93c8df4806c249c16e94cf87ad779b7ee37be0028d6057fbe4f4d56" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070204566024452 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070204567017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070173540016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070173540015457 5ustar corecore